optimal control of odes and daes

243
8/19/2019 Optimal Control of ODES and DAES http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 1/243 MATTHIAS GERDTS Optimal Control of Ordinary Differential Equations and Differential-Algebraic Equations Habilitation Thesis Department of Mathematics University of Bayreuth Accepted: December 2006 Revised Version: 23rd January 2007

Upload: andresmilquez

Post on 07-Jul-2018

222 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 1/243

MATTHIAS GERDTS

Optimal Control of Ordinary Differential Equations and

Differential-Algebraic Equations

Habilitation Thesis

Department of Mathematics

University of Bayreuth

Accepted: December 2006

Revised Version: 23rd January 2007

Page 2: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 2/243

Address of the Author:

Matthias Gerdts

Schwerpunkt Optimierung und ApproximationDepartment MathematikUniversitat Hamburg

D-20146 Hamburg

E-Mail: [email protected]

WWW: www.math.uni-hamburg.de/home/gerdts/

Copyright c

2006 by Matthias Gerdts

Page 3: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 3/243

Preface

This work was mainly developped at the Chair of Applied Mathematics at the Department of Mathematics of the University of Bayreuth. I am indebted to Prof. Dr. Frank Lempio for hisongoing support and many fruitful and valuable discussions. Particularly, I like to thank him forthe opportunity of giving a course on nondifferentiable optimization and the close collaborationin the course on discrete dynamic optimization in Borovets/Bulgaria. From both I benefited alot. In addition, he provided a very good atmosphere and working environment at his chair.Furthermore, I would like to thank Prof. Dr. Hans Josef Pesch for initiating my interest inoptimal control problems, the arrangement of my stay at the University of California, SanDiego, his encouragement and the motivation to finish this work.I thank Prof. Dr. Christof Buskens for the very close and friendly cooperation and for the

invitation to the University of Bremen for an interesting research stay.Last but not least I would like to express my thanks to all members of the Chair of AppliedMathematics and the Chair of Mathematics in the Engineering Sciences at the University of Bayreuth for the very pleasant and enjoyable time.Finally, I thank my parents for their steady support.

I dedicate this work to Katja and thank her for being there for me!

3

Page 4: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 4/243

Page 5: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 5/243

Contents

1 Introduction 1

2 Basics from Functional Analysis 13

2.1 Vector Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

2.2 Mappings, Dual Spaces, and Properties . . . . . . . . . . . . . . . . . . . . . . . 15

2.3 Function Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.4 Stieltjes Integral . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

2.5 Set Arithmetic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

2.6 Separation Theorems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252.7 Derivatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.8 Variational Equalities and Inequalities . . . . . . . . . . . . . . . . . . . . . . . . 29

3 Infinite and Finite Dimensional Optimization Problems 33

3.1 Problem Classes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

3.2 Existence of a Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

3.3 Conical Approximation of Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

3.4 First Order Necessary Conditions of Fritz-John Type . . . . . . . . . . . . . . . . 37

3.5 Constraint Qualifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

3.6 Necessary and Sufficient Conditions in Finte Dimensions . . . . . . . . . . . . . . 43

3.7 Perturbed Nonlinear Optimization Problems . . . . . . . . . . . . . . . . . . . . . 49

3.8 Numerical Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

3.9 Duality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70

3.10 Mixed-Integer Nonlinear Programs and Branch&Bound . . . . . . . . . . . . . . 75

4 Local Minimum Principles 81

4.1 Local Minimum Principles for Index-2 Problems . . . . . . . . . . . . . . . . . . 82

4.2 Local Minimum Principles for Index-1 Problems . . . . . . . . . . . . . . . . . . 110

5 Discretization Methods for ODEs and DAEs 115

5.1 General Discretization Theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1155.2 Backward Differentiation Formulae (BDF) . . . . . . . . . . . . . . . . . . . . . . 120

5.3 Implicit Runge-Kutta Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121

5.4 Linearized Implicit Runge-Kutta Methods . . . . . . . . . . . . . . . . . . . . . . 123

6 Discretization of Optimal Control Problems 137

6.1 Direct Discretization Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138

6.2 Calculation of Gradients . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148

6.3 Numerical Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160

6.4 Discrete Minimum Principle and Approximation of Adjoints . . . . . . . . . . . . 163

6.5 Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 181

i

Page 6: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 6/243

7 Selected Applications and Extensions 1857.1 Mixed-Integer Optimal Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1857.2 Open-Loop-Real-Time Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2017.3 Dynamic Parameter Identification . . . . . . . . . . . . . . . . . . . . . . . . . . 213

Bibliography 219

Page 7: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 7/243

Notation

Rn n-dimensional Euclidian space with norm · 2

0n zero in Rn

[t0, tf ] compact time interval in R with fixed t0 < tf

GN grid with N subintervalsX ,Y,Z Banach spaces · X norm on a Banach space X ΘX zero of a Banach space X Θ generic zero element of some space·, ·X scalar product on a pre-Hilbert space X U ε(x0) := x ∈ X | x − x0X < ε open ball around x0 with radius ε > 0

im(T ) := T (x) | x ∈ X image of X under the map T : X → Y ker(T ) := x ∈ X | T (x) = ΘX kernel or null space of a linear map T : X → Y T −1(S ) := x ∈ X | T (x) ∈ S preimage of the set S ⊆ Y of a map T : X → Y X ∗ topological dual space of X f X ∗ = supxX≤1 |f (x)| norm on X ∗ defining the strong topology

T ∗ adjoint operator of the linear map T : X → Y L(X, Y ) set of all linear, continuous mappings f : X → Y X/L factor space or quotient spaceL p([t0, tf ],R

n) space of all mappings f : [t0, tf ] → Rn that are boundedin the norm · p

W q,p([t0, tf ],Rn) Sobolev space of all absolutely continuous functions

f : [t0, tf ] → Rn that are bounded in the norm · q,p

C ([t0, tf ],Rn) space of continuous functions f : [t0, tf ] → Rn

AC ([t0, tf ],Rn) space of absolutely continuous functions f : [t0, tf ] →Rn

BV ([t0, tf ],Rn) functions of bounded variation

N BV ([t0, tf ],Rn) normalized functions of bounded variation tf

t0f (t)dµ(t) Riemann-Stieltjes integral

conv(S ) convex hull of the set S relint(S ) relative interior of the set S cone(S, x) conical hull of the set S − xcl(S ) closure of S int(S ) topological interior of S

S +, S − positive and negative dual cone of S F (x)(d) Frechet derivative of F at x in direction d∇F (x) = F (x) gradient of F : Rn → R

F x(x, y) partial Frechet derivative of F at (x, y)F (x; d) directional derivative of F at x in direction dδF (x)(d) Gateaux derivative of F at x in direction d∂F (x) Clarke’s subdifferential of F at x∂ BF (x) Bouligand-differential of F at x

iii

Page 8: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 8/243

lev(f, z) := x ∈ X | f (x) ≤ z level set of the function f at level zT (Σ, x) (sequential) tangent cone of Σ at x ∈ ΣT lin(K,S,x) linearizing cone of K and S at xT C (x) critical cone at x

H Hamilton function of an optimal control problem

L Lagrange function

Page 9: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 9/243

Chapter 1

Introduction

Historically, optimal control problems evolved from variational problems. Variational problemshave been investigated more thoroughly since 1696, although the first variational problems, e.g.queen Dido’s problem, were formulated in the ancient world already. In 1696 Johann Bernoulli(1667–1748) posed the Brachistochrone problem to other famous contemporary mathematicianslike Sir Isaac Newton (1643–1727), Gottfried Wilhelm Leibniz (1646–1716), Jacob Bernoulli(1654–1705), Guillaume Francois Antoine Marquis de L’Hospital (1661–1704), and EhrenfriedWalter von Tschirnhaus (1651–1708). Each of these distinguished mathematicians were able

to solve the problem. An interesting description of the Brachistochrone problem with manyadditional historical remarks as well as the solution approach of Johann Bernoulli exploitingFermat’s principle can be found in Pesch [Pes02].

Optimal control problems generalize variational problems by separating control and state vari-ables and admitting control constraints. Strongly motivated by military applications, optimalcontrol theory as well as solution methods evolve rapidly since 1950. The decisive break-through was achieved by the Russian mathematician Lev S. Pontryagin (1908–1988) and hiscoworkers V. G. Boltyanskii, R. V. Gamkrelidze, and E. F. Mishchenko in proving the max-imum principle, which provides necessary optimality conditions for optimal control problems,cf [PBGM64]. Almost at the same time also Magnus R. Hestenes [Hes66] proved a similartheorem. Since then many contributions in view of necessary conditions, cf., e.g., Jacobsen et

al. [JLS71], Girsanov [Gir72], Knobloch [Kno75], Neustadt [Neu76], Kirsch et al. [KWW78],Ioffe and Tihomirov [IT79], Maurer [Mau77, Mau79], Hartl et al. [HSV95], sufficient con-ditions, cf., e.g., Maurer [Mau81], Zeidan [Zei94], Malanowski [Mal97], Maurer and Picken-hain [MP95b], Maurer and Oberle [MO02], Malanowski et al. [MMP04], sensitivity analysis andreal-time optimal control, cf., e.g., Maurer and Augustin [MA01], Augustin and Maurer [AM01],Buskens and Maurer [BM96, BM01a], Malanowski and Maurer [MM96, MM98, MM01], Mau-rer and Pesch [MP94a, MP94b, MP95a], Pesch [Pes78, Pes79, Pes89a, Pes89b], and numeri-cal treatment, cf., e.g., Bulirsch [Bul71], Deuflhard [Deu74, Deu79], Deuflhard et al. [DPR76],Bryson and Ho [BH75], Diekhoff et al. [DLO+77], Chernousko and Lyubushin [CL82], Bockand Plitt [BP84], Kraft [KE85], Oberle [Obe86, OG01], Hargraves and Paris [HP87], Teo andGoh [TG87], Pesch and Bulirsch [PB94], Betts [Bet90], Hiltmann [HCB93], von Stryk [vS94],

Malanowski et al. [MBM97], Buskens [Bus98], Dontchev et al. [DHM00], Dontchev et al. [DHV00]have been released.

Applications of optimal control problems can be found in nearly any discipline, for example innatural sciences, engineering sciences, and economy.

While at the beginning the main focus was on optimal control problems subject to ordinarydifferential equations (ODEs), in the meanwhile also optimal control problems subject to partialdifferential equations (PDEs), differential-algebraic equations (DAEs), and stochastic optimalcontrol problems are under investigation.

In this book we will particularly focus on optimal control problems subject to differential-algebraic equations. DAEs are composite systems of differential equations and algebraic equa-

1

Page 10: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 10/243

2 CHAPTER 1. INTRODUCTION

tions and often are viewed as differential equations on manifolds. The implicit ODE

F (t, z(t), z(t), u(t)) = 0nz , t ∈ [t0, tf ] (1.1)

provides an example of DAEs. Herein, [t0, tf ] ⊂ R, t0 < tf is a given compact time interval and

F : [t0, tf ] × R

nz

× R

nz

× R

nu

→ R

nz

is a sufficiently smooth mapping. If the partial derivativeF z(·) in (1.1) happens to be non-singular, then (1.1) is just an ODE in implicit form and theimplicit function theorem allows to solve (1.1) for z in order to obtain an explicit ODE of typez(t) = f (t, z(t), u(t)). Hence, explicit or implicit ODEs are special cases of DAEs. The moreinteresting case occurs if the partial derivative F z(·) is singular . In this case (1.1) cannot besolved directly for z and (1.1) includes differential equations and algebraic equations at the sametime. Equations of such type are discussed intensively since the early 1970ies. Though at a firstglance DAEs seem to be very similar to ODEs they possess different solution properties, cf.Petzold [Pet82b]. Particularly, DAEs possess different stability properties compared to ODEs(see the perturbation index below) and initial values z(t0) = z0 have to be defined properly toguarantee at least locally unique solutions (see consistent initial values below).

It is important to mention, that, presently, (1.1) is too general and therefore too challenging tobeing tackled theoretically or numerically. Whenever necessary we will restrict the problem tomore simple problems. Most often, we will discuss so-called semi-explicit DAEs

x(t) = f (t, x(t), y(t), u(t)), (1.2)

0ny = g(t, x(t), y(t), u(t)), (1.3)

where the state z in (1.1) is decomposed into components x : [t0, tf ] → Rnx and y : [t0, tf ] → Rny .Herein, x(·) is referred to as differential variable and y(·) as algebraic variable . Correspondingly,(1.2) is called differential equation and (1.3) algebraic equation . Actually, the restriction to semi-explicit DAEs is (at least theoretically) not essential, since the implicit DAE (1.1) is equivalentwith the semi-explicit DAE

z(t) = y(t),

0nz = F (t, z(t), y(t), u(t)).

The functions x(·) and y(·) define the state of the upcoming optimal control problem. Thefunction u : [t0, tf ] → Rnu is considered as an external input and it is referred to as control variable . Usually, u(t) is restricted to a set U ⊆ Rnu , i.e.

u(t) ∈ U . (1.4)

Furthermore, the state and control variables may be restricted by mixed control-state constraints

c(t, x(t), y(t), u(t)) ≤ 0nc, (1.5)

pure state constraints s(t, x(t)) ≤ 0ns , (1.6)

and boundary conditions ψ(x(t0), x(tf )) = 0nψ

. (1.7)

Together with the objective function

ϕ(x(t0), x(tf )) +

tf

t0

f 0(t, x(t), y(t), u(t))dt (1.8)

we will investigate

c 2006 by M. Gerdts

Page 11: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 11/243

3

Problem 1.1 (DAE optimal control problem)Find functions x(·), y(·), and u(·) such that (1.8) is minimized subject to the constraints (1.2)– (1.7).

Herein, c : [t0, tf ]

×Rnx

×Rny

×Rnu

→ Rnc, s : [t0, tf ]

×Rnx

→ Rns , ψ : Rnx

×Rnx

→ Rnψ ,

ϕ : Rnx ×Rnx → R, and f 0 : [t0, tf ] × Rnx × Rny × Rnu → R are sufficiently smooth functions.Interestingly, the evaluation of the local minimum principle for optimal control problems subjectto explicit ODEs in absence of state and control constraints, i.e. nc = ns = ny = 0, U = Rnu,leads to a semi-explicit DAE itself, cf. Ioffe and Tihomirov [IT79]:

x(t) = f (t, x(t), u(t)),

λ(t) = −Hx(t, x(t), u(t), λ(t), l0),

0nu = Hu(t, x(t), u(t), λ(t), l0).

Herein, H(t,x,u,λ,l0) := l0f 0(t,x,u) + λf (t,x,u) denotes the Hamilton function. The state xand the adjoint (co-state) λ are differential variables while u is an algebraic variable.

For (1.2)-(1.3) it is not clear why u is viewed as a control variable while y should be a componentof the state. In fact, this assignment is somehow arbitrary from the optimal control theoreticpoint of view, since both, u and y, can be viewed as control variables. This becomes clearerin the chapter on local minimum principles for DAE optimal control problems, where it canbe seen, that both functions have very similar properties. Actually, this is the reason why thefunctions ϕ, ψ, and s only depend on x and not on y and u. But, recall that the DAE in realityis a model for, e.g., a robot, a car, an electric circuit, or a power plant. Hence, the DAE has ameaning for itself independent of whether it occurs in an optimal control problem or not. In thiscontext, it is necessary to distinguish between control u and algebraic variable y . The essentialdifference is, that an operator can choose the control u, whereas the algebraic variable y cannotbe controlled directly since it results from the state component x and the input u. For instance,

in the context of mechanical multi-body systems the algebraic variable y corresponds physicallyto a constraint force.The fact that y is defined by x and u, is mathematically taken into account for by imposingadditional regularity assumptions. The degree of regularity of the DAE (1.2)-(1.3) is measuredby a quantity called index . For specially structured DAEs, e.g. Hessenberg DAEs, the indexdefinitions coincide, whereas for general systems there are several different index definitions, cf.Duff and Gear [DG86a], Gear [Gea88, Gea90], Campbell and Gear [CG95], Hairer et al. [HLR89],and Marz [Mar95, Mar98b, Mar98a].

The perturbation index indicates the influence of perturbations and their derivatives on thesolution and therefore addresses the stability of DAEs .

Definition 1.2 (Perturbation Index, Hairer et al. [HLR89])The DAE (1.1) has perturbation index p ∈ N along a solution z on [t0, tf ], if p ∈ N is the smallest number such that for all functions z satisfying the perturbed DAE

F (t, z(t), ˙z(t), u(t)) = δ (t), (1.9)

there exists a constant S depending on F and tf − t0 with

z(t) − z(t) ≤ S

z(t0) − z(t0) + max

t0≤τ ≤tδ (τ ) + . . . + max

0≤τ ≤tδ ( p−1)(τ )

(1.10)

for all t ∈

[t0, tf ] whenever the expression on the right is less than or equal to a given bound.

c 2006 by M. Gerdts

Page 12: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 12/243

4 CHAPTER 1. INTRODUCTION

The perturbation index is p = 0, if the estimate

z(t) − z(t) ≤ S

z(t0) − z(t0) + max

t0≤τ ≤tf

τ

t0

δ (s)ds

(1.11)

holds.In the sequel we will make use of the subsequent lemma.

Lemma 1.3 (Gronwall) Let w, z : [t0, tf ] → R be integrable functions with

w(t) ≤ L

t

t0

w(τ )dτ + z(t)

for a.e. t ∈ [t0, tf ] with some constant L ≥ 0. Then it holds

w(t) ≤ z(t) + L

t

t0

exp(L(t − τ )) z(τ )dτ

for a.e. t ∈ [t0, tf ]. If z in addition belongs to L∞([t0, tf ],R) then it holds

w(t) ≤ z(·)∞ exp (L(t − t0))

for a.e. t ∈ [t0, tf ].

Proof. According to the assumption we may write

w(t) = a(t) + z(t) + δ (t)

with the absolutely continuous function

a(t) := L t

t0

w(τ )dτ

and a non-negative function δ (·) ∈ L1([t0, tf ],R). Introducing the expression for w in a yields

a(t) = L

t

t0

a(τ )dτ + L

t

t0

(z(τ ) + δ (τ )) dτ.

Hence, a solves the inhomogeneous linear differential equation

a(t) = La(t) + L (z(t) + δ (t))

for almost every t ∈ [t0, tf ] with initial value a(t0) = 0. The well-known solution formula forlinear differential equations yields

a(t) = L

t

t0

exp(L(t − τ )) (z(τ ) + δ (τ )) dτ

respectively

w(t) = L

t

t0

exp(L(t − τ )) (z(τ ) + δ (τ )) dτ + z(t) + δ (t).

Since δ (t) ≥ 0 the first assertion holds. If z is even essentially bounded we find

w(t) ≤ z(·)∞

1 + L

t

t0

exp(L(t − τ )) dτ

= z(·)∞ exp (L(t − t0)) .

c 2006 by M. Gerdts

Page 13: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 13/243

5

Remark 1.4 Obviously, the last assertion of Gronwall’s lemma remains true if we apply the norm · ∞ only to the interval [t0, t] instead of the whole interval [t0, tf ].

The ODE case:

To illustrate the perturbation index, we start with the initial value problem (IVP)

x(t) = f (t, x(t), u(t)), x(t0) = x0,

where f is assumed to be Lipschitz continuous w.r.t. x uniformly with respect to t.The (absolutely continuous) solution x can be written in integral form:

x(t) = x0 +

t

t0

f (τ, x(τ ), u(τ ))dτ.

Now, consider the perturbed IVP

˙x(t) = f (t, x(t), u(t)) + δ (t), x(t0) = x0

with an integrable perturbation δ : [t0, tf ] →Rnx and its solution

x(t) = x0 +

t

t0

(f (τ, x(τ ), u(τ )) + δ (τ )) dτ.

It holds

x(t) − x(t) ≤ x0 − x0 +

t

t0

f (τ, x(τ ), u(τ )) − f (τ, x(τ ), u(τ ))dτ +

t

t0

δ (τ )dτ

≤ x0 − x0 + L t

t0x(τ ) − x(τ )dτ + t

t0δ (τ )dτ .

Application of Gronwall’s lemma yields

x(t) − x(t) ≤

x0 − x0 + maxt0≤τ ≤t

τ

t0

δ (s)ds

exp(L(t − t0))

x0 − x0 + maxt0≤τ ≤tf

τ

t0

δ (s)ds

exp(L(tf − t0)) .

Hence, the ODE has perturbation index p = 0.

The index-1 case:

Consider the DAE (1.2)-(1.3) with solution (x, y) and initial value x(t0) = x0 and the perturbedIVP

˙x(t) = f (t, x(t), y(t), u(t)) + δ f (t), x(t0) = x0, (1.12)

0ny = g(t, x(t), y(t), u(t)) + δ g(t) (1.13)

in [t0, tf ]. Assume that

(i) f is Lipschitz continuous w.r.t. x and y with Lipschitz constant L f uniformly with respectto t;

c 2006 by M. Gerdts

Page 14: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 14/243

6 CHAPTER 1. INTRODUCTION

(ii) g is continuously differentiable and g y(t,x,y,u) is non-singular and bounded for all (t,x,y,u) ∈[t0, tf ] ×Rnx ×Rny × U .

According to (ii) the equation0ny = g(t,x,y,u) + δ g

can be solved for y ∈ Rny for any t ∈ [t0, tf ], x ∈ Rnx, u ∈ U by the implicit function theorem:

y = Y (t,x,u,δ g), g(t,x,Y (t,x,u,δ g), u) + δ g = 0ny .

Furthermore, Y is locally Lipschitz continuous w.r.t. t ,x, u, δ g with Lipschitz constant LY . Let(x(·), y(·)) denote the unperturbed solution of (1.2)-(1.3) resp. of (1.12)-(1.13) for δ f ≡ 0, δ g ≡ 0and (x(·), y(·)) the perturbed solution of (1.12)-(1.13). For the algebraic variables we get theestimate

y(t) − y(t) = Y (t, x(t), u(t), 0ny ) − Y (t, x(t), u(t), δ g (t))≤ LY (x(t) − x(t) + δ g(t)) .

With (i) we obtain

x(t) − x(t) ≤ x0 − x0 +

t

t0

f (τ, x(τ ), y(τ ), u(τ )) − f (τ, x(τ ), y(τ ), u(τ ))dτ

+

t

t0

δ f (τ )dτ

≤ x0 − x0 + Lf

t

t0

x(τ ) − x(τ ) + y(τ ) − y(τ )dτ +

t

t0

δ f (τ )dτ

≤ x0 − x0 + Lf (1 + LY )

t

t0

x(τ ) − x(τ )dτ + Lf

t

t0

δ g(τ )dτ

+ t

t0

δ f (τ )dτ .

Using the same arguments as in the ODE case we end up in the estimate

x(t) − x(t) ≤

x0 − x0 + (tf − t0)

Lf max

t0≤τ ≤tδ g(τ ) + max

t0≤τ ≤tδ f (τ )

·

· exp(Lf (1 + LY )(t − t0)) .

Hence, if the assumptions (i) and (ii) are valid, then the DAE (1.2)-(1.3) has perturbation index p = 1.

The index-k case:

The above procedure will work for a class of DAEs called Hessenberg DAEs.

Definition 1.5 (Hessenberg DAE)Let k ≥ 2. The semi-explicit DAE

x1(t) = f 1(t, x0(t), x1(t), x2(t), . . . , xk−2(t), xk−1(t), u(t)),

x2(t) = f 2(t, x1(t), x2(t), . . . , xk−2(t), xk−1(t), u(t)),

... . . .

xk−1(t) = f k−1(t, xk−2(t), xk−1(t), u(t)),

0ny

= g(t, xk−1

(t), u(t))

(1.14)

c 2006 by M. Gerdts

Page 15: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 15/243

7

is called Hessenberg DAE of order k. Herein, the differential variable is x = (x1, . . . , xk−1)

and the algebraic variable is y = x0.

For i = 1, . . . , k − 1 let f i be i times continuously differentiable and let u be sufficiently smooth.By (k

−1)-fold differentiation of the algebraic constraint, application of the implicit function

theorem and repeating the arguments for ODEs and index-1 DAEs it is easy to see that theHessenberg DAE of order k has perturbation index k, if the matrix

M := gxk−1(·) · f k−1,xk−2

(·) · · · f 2,x1(·) · f 1,x0(·) (1.15)

is non-singular.

Another index concept corresponds to the definition of the order of active state constraintsin optimal control problems, cf. Maurer [Mau79]. Notice, that the algebraic constraint (1.3)can be interpreted as a state resp. mixed control-state constraint in an appropriate optimalcontrol problem, which is active on [t0, tf ]. The idea is to differentiate the algebraic constraint(1.3) w.r.t. time and to replace the derivative x according to (1.2) until the resulting equations

can be solved for the derivative y by the implicit function theorem. The smallest number of differentiations needed is called differential index , c.f. Brenan et al. [BCP96].

We illustrate the differential index for (1.2)-(1.3) and assume

(i) g and u are continuously differentiable.

(ii) gy(t,x,y,u) is non-singular and gy(t,x,y,u)−1 is bounded for all (t,x,y,u) ∈ [t0, tf ] ×Rnx × Rny × U .

Differentiation of the algebraic constraint (1.3) w.r.t. time yields

0ny = gt[t] + gx[t]

· x(t) + gy[t]

· y(t) + gu[t]

· u(t)

= gt[t] + gx[t] · f [t] + gy[t] · y(t) + gu[t] · u(t),

where, e.g., gt[t] is an abbreviation for gt(t, x(t), y(t), u(t)). Since g y[t]−1 exists and is boundedwe can solve this equality for y and obtain the differential equation

y(t) = − gy[t]

−1 gt[t] + gx[t] · f [t] + gu[t] · u(t)

. (1.16)

Hence, it was necessary to differentiate the algebraic equation once in order to obtain a differ-ential equation for the algebraic variable y. Consequently, the differential index is one. Equa-tions (1.2) and (1.16) are called the underlying ODE of the DAE (1.2)-(1.3). Often, ODEs areviewed as index-0 DAEs, since the underlying ODE is identical with the ODE itself and no

differentiations are needed.In the same way it can be shown that the Hessenberg DAE of order k ≥ 2 has differential indexk, if the matrix M in (1.15) is non-singular.

Consequently, the differential index and the perturbation index coincide for Hessenberg DAEs,cf. also Campbell and Gear [CG95]. For more general DAEs the difference between perturbationindex and differential index can be arbitrarily high, cf. Hairer and Wanner [HW96], p. 461.

Remark 1.6 Notice, that the index calculations are questionable in the presence of a control variable u. We assumed that u is sufficiently smooth which is not the case for many optimal control problems. Nevertheless, often the optimal control u is at least piecewise smooth and hence the interpretation remains valid locally almost everywhere.

c 2006 by M. Gerdts

Page 16: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 16/243

8 CHAPTER 1. INTRODUCTION

Consistent initial values

DAEs not only differ in their stability behavior from explicit ODEs. Another difference is that

initial values have to be defined properly, that is they have to be consistent. We restrict thediscussion to Hessenberg DAEs of order k. A general definition of consistency can be found inBrenan et al. [BCP96].

Define

g(0)(t, xk−1(t), u(t)) := g(t, xk−1(t), u(t)). (1.17)

Of course, an initial value for xk−1 at t = t0 has to satisfy this equality. But this is notsufficient, because also time derivatives of (1.17) impose additional restrictions on initial values.This can be seen as follows. Differentiation of g(0) w.r.t. time and substitution of xk−1(t) =f k−1(t, xk−2(t), xk−1(t), u(t)) leads to the equation

0ny = gt[t] + gxk−1 [t] · f k−1(t, xk−2(t), xk−1(t), u(t)) + gu[t] · u(t)

=: g(1)(t, xk−2(t), xk−1(t), u(t), u(t)),

which has to be satisfied as well. Recursive application of this differentiation and substitutionprocess leads to the equations

0ny = g( j)(t, xk−1− j (t), . . . , xk−1(t), u(t), u(t), . . . , u( j)(t)), j = 1, 2, . . . , k − 2, (1.18)

and

0ny = g(k−1)(t, x0(t), x1(t), . . . , xk−1(t), u(t), u(t), . . . , u(k−1)(t)). (1.19)

Since the equations (1.18)-(1.19) do not occur explicitly in the original system (1.14), theseequations are called hidden constraints of the Hessenberg DAE. With this notation consistencyis defined as follows.

Definition 1.7 (Consistent Initial Value)

Let x(t0) = (x0(t0), x1(t0), . . . , xk−1(t0)) be given and let u be sufficiently smooth. x(t0) is called consistent for the Hessenberg DAE of order k, if x(t0) satisfies the equations (1.17)-(1.19) at t = t0. The differential component is called consistent if equations (1.17)-(1.18) are satisfied.

Remark 1.8 The dependence on derivatives of the control in (1.18) and (1.19) can be avoided by introducing additional state variables ξ

j, j = 0, . . . , k

−2, by

ξ 0 := u, ξ 1 := u, . . . ξ k−2 := u(k−2)

satisfying the differential equations

ξ 0 = ξ 1, . . . ξ k−3 = ξ k−2, ξ k−2 = u

and to consider u := u(k−1) as the new control, cf. M¨ uller [M¨ ul03]. Clearly, this approach is nothing else than constructing a sufficiently smooth control u for the original problem. The resulting problem is not equivalent to the original problem anymore. Nevertheless, this strategy is very useful from a practical point of view.

c 2006 by M. Gerdts

Page 17: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 17/243

9

Index reduction and stabilization

The perturbation index measures the stability of the DAE. DAEs with perturbation index greater

than or equal to two are called higher index DAEs . With increasing perturbation index the DAEsuffers from increasing ill-conditioning, since not only the perturbation δ occurs in (1.10) butalso derivatives thereof. Thus, even for small perturbations (in the supremum norm) largedeviations in the respective solutions are possible. Higher index DAEs occur, e.g., in mechanics,cf. Fuhrer [Fuh88], Simeon [Sim94], Arnold [Arn95], process system engineering, cf. Engl etal. [EKKvS99], and electrical engineering, cf. Marz and Tischendorf [MT97], Gunther [Gun95].To avoid the problems of severe ill-conditioning it is possible to reduce the index by replacingthe algebraic constraint in (1.14) by its derivative g (1). It is easy to see that the resulting DAEhas perturbation index k − 1. More generally, the algebraic constraint g(0)(·) = 0ny can be

replaced by any of the constraints g( j)(·) = 0ny with 1 ≤ j ≤ k − 1. Then, the resulting DAEhas perturbation index k

− j. The drawback of this index reduction approach is that numerical

integration methods suffer from the so-called drift-off effect . Since a numerical integrationscheme when applied to an index reduced DAE only obeys hidden constraints up to a certainlevel, the numerical solution will not satisfy the higher level constraints and it can be observedthat the magnitude of violation increases with time t. Hence, the numerical solution driftsoff the neglected constraints even when the initial value was consistent. Especially for longtime intervals the numerical solution may deviate substantially from the solution of the originalDAE. One possibility to avoid this drawback is to perform a projection step onto the neglectedconstraints after each successful integration step for the index reduced system, cf. Ascher andPetzold [AP91] and Eich [Eic93].Another approach is to stabilize the index reduced problem by adding the neglected constraintsto the index reduced DAE. The resulting system is an overdetermined DAE and numericalintegration methods have to be adapted, cf. Fuhrer [Fuh88], Fuhrer and Leimkuhler [FL91].The stabilization approach when applied to mechanical multi-body systems is equivalent withthe GGL-stabilization, cf. Gear et al. [GLG85].The above mentioned definitions are illustrated for a very important field of application – me-chanical multi-body systems.

Example 1.9 (Mechanical Multi-body Systems)Let a system of n rigid bodies be given. Every body of mass mi has three translational andthree rotary degrees of freedom. The position of body i in a fixed reference system is given by itsposition ri = (xi, yi, zi). The orientation of the bodies coordinate system with respect to thereference coordinate system is given by the angles α i, β i, and γ i. Hence, body i is characterized

by the coordinatesq i = (xi, yi, zi, αi, β i, γ i) ∈ R

6

and the whole multi-body system is described by

q = (q 1, . . . , q n) ∈ R6n.

In general the motion of the n bodies is restricted by holonomic constraints

0ng = g(q ).

The kinetic energy of the multi-body system is given by

T (q, q ) = 1

2

n

i=1 mi · q i q i + w

i · I i · wi ,

c 2006 by M. Gerdts

Page 18: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 18/243

10 CHAPTER 1. INTRODUCTION

where wi denotes the angular velocity of body i w.r.t. the reference system and I i is the momentof inertia of body i. The Euler-Lagrangian equations of the first kind are given by

d

dt T q(q, q )

T q(q, q )

= F (q, q, u) − g(q )λ,

0ng = g(q ),

where F is the vector of applied forces and moments, which may depend on the control u.Explicit calculation of the derivatives leads to the descriptor form of mechanical multi-bodysystems:

q (t) = v(t), (1.20)

M (q (t))v(t) = f (q (t), v(t), u(t)) − g(q (t))λ(t), (1.21)

0ng = g(q (t)), (1.22)

where M is the symmetric and positive definite mass matrix and f includes the applied forces

and moments and the Coriolis forces. Multiplication of the second equation with M −1

yields aHessenberg DAE of order 3.Let g be twice continuously differentiable. The constraint g(q (t)) = 0ng is called constraint on position level . Differentiation w.r.t. time of this algebraic constraint yields the constraint on velocity level

g(q (t)) · v(t) = 0ng

and the constraint on acceleration level

g(q (t)) · v(t) + gqq(q (t))(v(t), v(t)) = 0ng .

Replacing v by

v(t) = M (q (t))−1 f (q (t), v(t), u(t)) − g

(q (t))

λ(t)yields

g(q (t))M (q (t))−1

f (q (t), v(t), u(t)) − g(q (t))λ(t)

+ gqq (q (t))(v(t), v(t)) = 0ng .

If rank(g(q )) = ng, then the matrix g (q )M (q )−1g(q ) is non-singular and the latter equationcan be solved for the algebraic variable λ. By the same reasoning as before it can be shown thatthe descriptor form has index 3. Consistent initial values (q 0, v0, λ0) satisfy the constraints onposition, velocity, and acceleration level.For numerical methods it is advisable to perform an index reduction, i.e. the constraint onposition level is replaced by the constraint on velocity or acceleration level. An even better idea

is to use the GGL-stabilization

q (t) = v(t) − g(q (t))µ(t), (1.23)

M (q (t))v(t) = f (q (t), v(t), u(t)) − g(q (t))λ(t), (1.24)

0ng = g(q (t)), (1.25)

0ng = g(q (t)) · v(t). (1.26)

This DAE is equivalent to an index 2 Hessenberg DAE, if the second equation is multiplied withM −1. Furthermore, differentiation of the first algebraic equation yields

0ny = g(q (t))

· v(t)

−g(q (t))µ(t) =

−g(q (t))g(q (t))µ(t).

c 2006 by M. Gerdts

Page 19: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 19/243

11

If rank(g(q )) = ng then g(q )g(q ) is non-singular and it follows µ ≡ 0ng . Hence, the GGL-stabilization (1.23)-(1.26) is equivalent to the overdetermined (stabilized) descriptor form

q (t) = v(t),

M (q (t))v(t) = f (q (t), v(t), u(t))−

g(q (t))λ(t),

0ng = g(q (t)),

0ng = g(q (t)) · v(t).

After these introductory remarks on DAEs an outline of the book is as follows. In Chapter 2the functional analytic background needed for the derivation of local minimum principles forDAE optimal control problems is provided. For the readers convenience we included a ratherdetailed presentation of the background material. Although most of the material is standard, wefound that some parts, e.g. the more advanced properties of Stieltjes integrals and the resultson variational equalities and inequalities involving functions of bounded variation and Stieltjes

integrals, are hard to find in the literature.Chapter 3 summarizes results on finite and infinite optimization problems. Similar as in Chap-ter 2, for the sake of completeness, we provide a detailed presentation of the results includingmost of the proofs. We found it useful to include, e.g., the proof of the first order necessaryFritz-John conditions since it sheds light on many difficulties arising in infinite dimensions.Furthermore, the exposition attempts to combine several versions of necessary conditions fromthe literature, that differ sligthly in their assumptions or the problem classes under consider-ation, and to adapt them to our purposes. Readers who are familiar with infinite and finitedimensional optimization theory may skip large parts of this chapter. However, Chapter 3 isof central importance for all upcoming theoretical and numerical approaches towards optimalcontrol problems.

Since most current numerical integration methods are only capable of solving nonlinear index-1and index-2 Hessenberg DAEs, we restrict the discussion to optimal control problems subjectto these two classes of DAEs. Of course, explicit ODEs are included in either of such DAEs assubclasses. The optimal control problem is considered as an infinite dimensional optimizationproblem of type

F (x,y,u) → min s.t. G(x,y ,u) ∈ K, H (x,y ,u) = Θ, u ∈ U ,where F, G, H are mappings between appropriate Banach spaces, K is a cone, and U is a convexset with non-empty interior. The proof of the local minimum principles in Chapter 4 exploitsfirst order necessary conditions for such general infinite dimensional optimization problems.An additional regularity assumption similar to the Mangasarian-Fromowitz condition in finitedimensional optimization provides a constraint qualification for the local solution. This chapteris a major contribution of the author to the theory of optimal control processes with DAEconstraints. For such problems only very few theoretical results are known up to now.The local minimum principle usually leads to a boundary value problem and thus providesthe basis of the so-called indirect solution approach for optimal control problems. The indirectmethod attempts to solve the minimum principle resp. the boundary value problem numerically.In Chapter 3 we summarize also well-known necessary and sufficient conditions for finite dimen-sional optimization problems. These conditions are the basis for the direct solution approachfor optimal control problems. This direct approach is based on a discretization of the optimalcontrol problem and leads to a finite dimensional optimization problem, which is solved numer-ically by the sequential quadratic programming (SQP) method. Direct discretization methods

c 2006 by M. Gerdts

Page 20: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 20/243

12 CHAPTER 1. INTRODUCTION

for optimal control problems are discussed in Chapter 6. These methods need integration meth-ods for DAEs that are briefly summarized in Chapter 5. In particular, the class of linearizedRunge-Kutta methods is introduced and analysed. This new class turns out to be efficient fordiscretized optimal control problems. Application of the necessary finite dimensional Fritz-Johnconditions yields a discrete version of the local minimum principle and allows to compute ap-proximations for the adjoints (co-states). On the other hand, the adjoint equation can be usedfor the calculation of gradients within the SQP method. An alternative method for calculatinggradients is the sensitivity equation approach, which is preferable if the number of variablesis small compared to the number of constraints. Finally, selected applications from real-timeoptimization, parameter identification, and mixed-integer optimal control are discussed in Chap-ter 7. The methods and results presented in chapters 6 and 7 include contributions of the authorto the numerical aspects of DAE optimal control problems, e.g. sensitivity analysis, gradientcalculation, determination of consistent initial values and mixed-integer optimal control.

c 2006 by M. Gerdts

Page 21: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 21/243

Page 22: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 22/243

14 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

In order to define a topology in a vector space it is sufficient to specify a basis A of open setsaround zero. A basis around zero is characterized by the property that for any neighborhoodV of zero there exists U ∈ A such that U ⊆ V . Herein, every open set containing x is called a

neighborhood of x. Then, a set S is open, if and only if for every x ∈ S there exists an elementU

∈ A such that x + U

⊆ S . Furthermore, x

∈ S is called an interior point of S , if there is a

neighborhood U of x with U ⊆ S . The set of all interior points of S is denoted by int(S ). Theclosure cl(S ) of S is the set of all points x satisfying U ∩ S = ∅ for all neighborhoods U of x.x is a boundary point of S , if x ∈ cl(S ) and x ∈ int(S ). We have that S is open, if and only if S = int(S ). Furthermore, S is closed, if and only if S = cl(S ).A metric space is a tupel (X, d), where X is a set and d : X × X → R is a mapping such thatfor every x, y,z ∈ X it holds

(i) d(x, y) ≥ 0 and d(x, y) = 0, if and only if x = y;

(ii) d(x, y) = d(y, x);

(iii) d(x, y) ≤ d(x, z) + d(z, y).

d is called a metric on X .In a metric space the sequence xn is said to converge to x ∈ X , that is xn → x, if d(xn, x) → 0.If a sequence converges, its limit is unique. Recall that a metric space X is called complete , if every Cauchy sequence from X has a limit in X . A sequence xn in X is a Cauchy sequence ,if for every ε > 0 there is a N (ε) ∈ N such that d(xn, xm) < ε for all n,m > N (ε). In a metricspace every convergent sequence is a Cauchy sequence.A metric space (X, d) is called compact , if for every sequence xn in X there exists a subsequencexnk

converging to an element x ∈ X . Every compact metric space is complete. In finitedimensional spaces, compactness is equivalent with closedness and boundedness. In infinitedimensional spaces this is not true. For instance, it is shown in Luenberger [Lue69], p. 40, thatthe unit sphere in general is not compact. A subset S of a metric space (X, d) is called dense in X , if cl(S ) = X holds.Let X be a vector space over K. The tupel (X, · X ) is called a normed vector space , if · X : X → R is a mapping such that for every x, y ∈ X and every λ ∈ K it holds

(i) xX ≥ 0 and xX = 0 if and only if x = ΘX ;

(ii) λxX = |λ| · xX ;

(iii) x + yX ≤ xX + yX .

The mapping · X is called a norm on X . Since every norm defines a metric by d(x, y) :=x − yX the terminologies ‘convergence, complete, Cauchy sequence,. . . ’ can be translateddirectly to normed spaces (X,

· X ).

Definition 2.1.1 (Banach space)A complete normed vector space is called Banach space.

In a Banach space for r > 0 let

U r(x) := y ∈ X | y − xX < r,

U r(x) := y ∈ X | y − xX ≤ rdenote the open and closed balls around x with radius r, respectively. Then A = U r(ΘX ) | r >0 defines a basis of open sets about zero. We say, the norm · X induces the strong topology on X .Finally, we introduce Hilbert spaces.

c 2006 by M. Gerdts

Page 23: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 23/243

2.2. MAPPINGS, DUAL SPACES, AND PROPERTIES 15

Definition 2.1.2 (Inner Product, Scalar Product)Let X be a vector space over K = R or K = C. The mapping ·, ·X : X × X → K is called an inner product or scalar product, if the following conditions hold for all x,y,z ∈ X and all λ ∈ K.

(i) x, yX = y, xX ;

(ii) x + y, zX = x, zX + y, zX ;

(iii) x,λyX = λx, yX ;

(iv) x, xX ≥ 0 and x, xX = 0, if and only if x = ΘX .

Definition 2.1.3 (pre-Hilbert Space, Hilbert Space)

A pre-Hilbert space is a vector space together with an inner product. A complete pre-Hilbert space is called Hilbert space.

Notice, that

x

X := x, x

X defines a norm on a pre-Hilbert space X .

2.2 Mappings, Dual Spaces, and Properties

From now on, unless otherwise specified, (X, · X ), (Y, · Y ), (Z, · Z ), . . . denote real Banachspaces over K = R with norms · X , · Y , · Z , . . .. If no confusion is possible, we omit thenorms and call X, Y , Z , . . . Banach spaces and assume that appropriate norms are defined.

Definition 2.2.1Let T : X → Y be a mapping from a Banach space (X, · X ) into a Banach space (Y, · Y ).

(i) The image of T is given by im (T ) := T (x) | x ∈ X .

The kernel of T is given by

ker (T ) := x ∈ X | T (x) = ΘY .

Given a set S ⊆ Y , the preimage of S under T is given by

T −1(S ) := x ∈ X | T (x) ∈ S .

(ii) T is called linear, if

T (x1 + x2) = T (x1) + T (x2), T (λx1) = λT (x1) ∀x1, x2 ∈ X, λ ∈ R.

(iii) T is called continuous at x

∈ X , if for every sequence xi

→x it holds T (xi)

→T (x). T is

called continuous on X , if T is continuous at all x ∈ X .

(iv) Let D ⊆ X be open. A function f : D → Y is called locally Lipschitz continuous at x ∈ Dwith constant L, if there exists a ε > 0 such that

f (y) − f (z)Y ≤ Ly − zX ∀y, z ∈ U ε(x).

(v) T is called bounded, if T (x)Y ≤ C · xX holds for all x ∈ X and some constant C ≥ 0.

The operator norm is defined by

T X,Y = supx=ΘX

T (x)Y

x

X

= supxX≤1

T (x)Y = supxX=1

T (x)Y .

c 2006 by M. Gerdts

Page 24: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 24/243

16 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

(vi) If Y = R, then T is called a functional.

(vii) A functional f : X → R is called upper semicontinuous at x, if for every sequence xiwith xi → x it holds

limsup

i→∞

f (xi)

≤ f (x).

A functional f : X → R is called lower semicontinuous at x, if for every sequence xiwith xi → x it holds

f (x) ≤ lim inf i→∞

f (xi).

(viii) The set of all linear continuous functionals on X endowed with the norm

f X ∗ = supxX≤1

|f (x)|

is called dual space of X and is denoted by X ∗. This norm defines the strong topology onX ∗. X is called reflexive, if (X ∗)∗ = X holds with respect to the strong topology.

(ix) Let T : X → Y be linear. The adjoint operator T ∗ : Y ∗ → X ∗ is a linear operator defined by

T ∗(y∗)(x) = y∗(T (x)) ∀y∗ ∈ Y ∗, x ∈ X.

A linear operator is continuous, if and only if it is bounded. A linear operator is continuous, if and only if it is continuous at zero. If T is continuous, then so is T ∗. The set L(X, Y ) of alllinear continuous mappings from X into Y endowed with the operator norm ·X,Y is a Banachspace. The dual space of a Banach space is a Banach space.

Theorem 2.2.2 (cf. Ljusternik and Sobolew [LS76], p. 109, Th. 4)Let X, Y be Banach spaces and T : X

→ Y a continuous, linear, surjective, and injective

operator. Then the inverse operator T −1 exists and it is linear and continuous.

It is well-known, that given n Banach spaces X 1, X 2, . . . , X n with norms · 1, · 2, . . . , · n,the product space

X = X 1 × X 2 × · · · × X n

equipped with one of the norms

xX = max1≤i≤n

xiX i , xX =

ni=1

xi pX i

1/p

, p ∈ N,

is also a Banach space. The dual space of X is

X ∗ = x∗ = (x∗1, x∗2, . . . , x∗n) | x∗i ∈ X ∗i , i = 1, . . . , n, x∗(x) =n

i=1

x∗i (xi).

The following theorem can be found in Werner [Wer95], p. 135, Th. IV.3.3.

Theorem 2.2.3 (Open Mapping Theorem)Let T : X → Y be a linear, continuous, and surjective operator. Let S ⊆ X be open. Then T (S ) ⊆ Y is open.

Let T : X → Y be linear and continuous. Then, the following statements are equivalent, cf.Werner [Wer95], p. 143, Th. IV.5.1:

c 2006 by M. Gerdts

Page 25: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 25/243

2.2. MAPPINGS, DUAL SPACES, AND PROPERTIES 17

(i) im(T ) is closed.

(ii) im(T ) = (ker(T ∗))⊥ := x ∈ X | x∗(x) = 0 ∀x∗ ∈ X ∗.

(iii) im(T ∗) is closed.

(iv) im(T ∗) = (ker(T ))⊥ := x∗ ∈ X ∗ | x∗(x) = 0 ∀x ∈ X .

Let T : X → Y be a linear and continuous mapping. Then, ker(T ) is a closed subspace of X .

Definition 2.2.4 (Factor space, Quotient space, codimension)Let X be a vector space and L ⊆ X a subspace. The factor space or quotient space X/L consists of all sets [x], x ∈ X , where [x] := x + L. Vectors x, y ∈ X are called L-equivalent, if x − y ∈ L.The dimension of X/L is called codimension of L or defect of L.

If X is finite dimensional, then dim(X/L) = dim(X ) − dim(L), cf. Kowalsky [Kow69], p. 214.

Let L be a linear subspace of the real vector space X and x

∈ X . The set x + L is called a

linear manifold . Let eii∈I be a basis of X and e j j∈J , J ⊆ I a basis of L. The dimension of the linear manifold x + L is defined as |J | and the codimension or defect of x + L is defined as|I \ J |. Hyperplanes are linear manifolds having defect one. Furthermore, H is a hyperplane,if and only if it can be represented as H = x ∈ X | f (x) = γ , where f is a non-zero linearfunctional f : X → R and γ ∈ R. Given a set S in a real vector space X , the codimension of S is defined as the codimension of the linear subspace parallel to the affine hull of S .

If L is a closed subspace of the Banach space X then

[x]X/L := inf y − xX | y ∈ L

defines a norm on X/L and X/L endowed with the norm · X/L is a Banach space.

The canonical mapping w : X → X/L, x → [x] is surjective and ker(w) = L, cf. Kowal-sky [Kow69], p. 213.

Theorem 2.2.5 (cf. Kowalsky [Kow69], p. 214)Let T : X → Y be a linear mapping from the vector space X into the vector space Y and L ⊆ ker (T ) a subspace of X . Then T can be factorized uniquely as T = T w with a linear mapping T : X/L → Y . Furthermore, T is injective, if and only if ker (T ) = L. T is surjective,if and only if T is surjective.

As a direct consequence of the theorem we find

Corollary 2.2.6 Let T : X → Y be a linear and surjective mapping from the vector space X onto the vector space Y . Then there exists a linear, surjective and injective mapping T :X/ker (T ) → Y .

A linear continuous operator T : X → Y is called an isomorphism , if T is surjective and injectiveand T −1 is linear and continuous. Two normed spaces X and Y are called isomorph , if thereexists an isomorphism between X and Y .

Theorem 2.2.7 Let X, Y be Banach spaces and T : X → Y a linear, continuous, and sur- jective operator. Then there exists a linear, continuous, surjective, and injective mapping T : X/ker (T ) → Y such that T −1 exists and it is linear and continuous (that is, X/ker (T )and Y are isomorphic).

c 2006 by M. Gerdts

Page 26: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 26/243

Page 27: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 27/243

2.3. FUNCTION SPACES 19

2.3 Function Spaces

Some important real Banach and Hilbert spaces are summarized in the sequel.

2.3.1 L p-Spaces

Let 1≤

p < ∞

. The space L p([a, b],R) consists of all measurable functions f : [a, b] →

R with b

a|f (t)| pdt < ∞,

where the integral denotes the Lebesgue integral. Notice, that functions that differ only on aset of measure zero, i.e. functions that are equal almost everywhere, are considered to be thesame. In this sense, the elements of L p are equivalence classes. The space L∞([a, b],R) consistsof all measurable functions f : [a, b] → R which are essentially bounded, i.e.

ess supa≤t≤b

|f (t)| := inf N ⊂[a,b]µ(N )=0

supt∈[a,b]\N

|f (t)| < ∞.

The spaces L p([a, b],R), 1

≤ p <

∞ endowed with the norm

f p :=

b

a|f (t)| pdt

1/p

and the space L∞([a, b],R) endowed with the norm

f ∞ := ess supa≤t≤b

|f (t)|

are Banach spaces, cf. Kufner et al. [KOS77], Ths. 2.8.2, 2.11.7. For 1 < p < ∞ the dual spaceof L p([a, b],R) is given by Lq([a, b],R) where 1/p + 1/q = 1, i.e. for f ∗ ∈ (L p([a, b],R))∗ thereexists a unique function g ∈ Lq([a, b],R) such that

f ∗

(f ) = b

a f (t)g(t)dt, ∀f ∈ L p

([a, b],R

),

and f ∗ = gq , cf. Kufner et al. [KOS77], Th. 2.9.5. For 1 < p < ∞ the spaces L p arereflexive, cf. Kufner et al. [KOS77], Th. 2.10.1.The dual space of L1([a, b],R) is given by L∞([a, b],R), i.e. for f ∗ ∈ (L1([a, b],R))∗ there existsa unique function g ∈ L∞([a, b],R) such that

f ∗(f ) =

b

af (t)g(t)dt, ∀f ∈ L1([a, b],R),

cf. Kufner et al. [KOS77], Th. 2.11.8. The spaces L1([a, b],R) and L∞([a, b],R) are not reflexive,cf. Kufner et al. [KOS77], Ths. 2.11.10, 2.11.11.

The dual space of L

([a, b],R

) does not have a nice structure. According to Kufner et al. [KOS77],Rem. 2.17.2, the space L∞([a, b],R) is isometrically isomorph with the space of all finitely addi-tive measures on the family of measurable subsets of [a, b] which are absolutely continuous w.r.t.the Lebesgue measure.L2([a, b],R) is a Hilbert space with the inner product

f, gL2 :=

b

af (t)g(t)dt.

For 1 ≤ p ≤ ∞ the space L p([a, b],Rn) is defined as the product space

L p([a, b],Rn) := L p([a, b],R) × · · · × L p([a, b],R),

where each element f of L p([a, b],Rn) is a mapping from [a, b] in Rn.

c 2006 by M. Gerdts

Page 28: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 28/243

20 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

2.3.2 Absolutely Continuous Functions

A function f : [a, b] → R is said to be absolutely continuous , if for every ε > 0 there exists aδ (ε) > 0 such that

mi=1

|bi − ai| < δ (ε) ⇒m

i=1|f (bi) − f (ai)| < ε,

where m ∈ N is arbitrary and (ai, bi) ⊆ [a, b], i = 1, . . . , m are disjoint intervals. The set of allabsolutely continuous functions is called AC ([a, b],R). The following properties of absolutelycontinuous functions can be found in Natanson [Nat75].If f ∈ L1([a, b],R) then the undetermined integral

F (t) := C +

t

af (τ )dτ

is absolutely continuous and it holds

F (t) = f (t) a.e. in [a, b].

On the other hand, if f ∈ L1([a, b],R) exists everywhere and is finite, then it holds

f (t) = f (a) +

t

af (τ )dτ. (2.3.1)

Furthermore, the partial integration b

af (t)g(t)dt +

b

ag(t)f (t)dt = [f (t)g(t)]b

a .

holds for absolutely continuous functions f and g on [a, b].

In addition, an absolutely continuous function on [a, b] is continuous, of bounded variation,f exists almost everywhere in [a, b] and satisfies (2.3.1) for a ≤ t ≤ b. If f is zero almosteverywhere in [a, b], then f is constant. If the derivatives f and g of absolutely continuousfunctions f and g are equal almost everywhere, then the difference f − g is constant.

2.3.3 W q,p-Spaces

Let 1 ≤ q, p ≤ ∞. The space W q,p([a, b],R) consists of all absolutely continuous functionsf : [a, b] →R with absolutely continuous derivatives up to order q − 1 and

f q,p < ∞,

where the norm is given by

f q,p :=

qi=0

f (i) p p

1/p

, 1 ≤ p < ∞,

f q,∞ := max0≤i≤q

f (i)∞.

The spaces W q,p([a, b],R), 1 ≤ q, p ≤ ∞ endowed with the norm · q,p are Banach spaces. Thespaces W q,2([a, b],R) are Hilbert spaces with the inner product

f, gW q,2 :=

q

i=0

b

af (i)(t)g(i)(t)dt.

c 2006 by M. Gerdts

Page 29: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 29/243

2.4. STIELTJES INTEGRAL 21

2.3.4 Functions of Bounded Variation

A function f : [a, b] → R is said to be of bounded variation , if there exists a constant K suchthat for any partition

Gm := a = t0 < t1 < .. . < tm = bof [a, b] it holds

mi=1

|f (ti) − f (ti−1)| ≤ K.

The total variation of f is

T V (f ,a,b) := supall Gm

mi=1

|f (ti) − f (ti−1)|.

The space BV ([a, b],R) consists of all functions of bounded variation on [a, b].

We summarize some facts about functions of bounded variation, cf. Natanson [Nat75].

• The derivative f of a function f ∈ BV ([a, b],R) exists almost everywhere in [a, b] and the

integral b

a f (t)dt exists.

• f ∈ BV ([a, b],R) possesses only countably many jumps. For every jump point of f theleft and right-sided limits exist.

• Every function f of bounded variation can be represented as

f (t) = g(t) + s(t) + r(t),

where g is absolutely continuous, s is the jump function of f , i.e.

s(a) = 0,

s(t) = (f (a+) − f (a)) +ti<t

(f (ti+) − f (ti−)) + (f (t) − f (t−)), a < t ≤ b.

and r is singular. Herein, r is called singular , if it is a non-constant, continuous functionof bounded variation, whose derivative is zero almost everywhere. Note, that a singularfunction is not absolutely continuous, since then it would be constant.

If f is continuous, then s is zero. If f is continuous and monotonically increasing, then gand r are also monotonically increasing.

2.4 Stieltjes IntegralThe Stieltjes integral generalizes the Riemann integral and is needed for the formulation of theminimum principle for state constrained optimal control problems. Let f and µ be two functionsdefined on the interval [a, b]. Let a partition

Gm := a = t0 < t1 < .. . < tm = b

of [a, b] be given. Let ξ i ∈ [ti−1, ti] be arbitrary and

S (f, µ) :=m

i=1

f (ξ i) (µ(ti) − µ(ti−1)) .

c 2006 by M. Gerdts

Page 30: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 30/243

22 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

If S (f, µ) converges to a finite value S independently of the points ξ i as maxiti+1 − ti → 0,then S is called Stieltjes integral of f w.r.t. µ and we write

b

a

f (t)dµ(t).

More precisely, S is called Stieltjes integral of f w.r.t. µ, if for every ε > 0 there exists a δ > 0such that for every partition with maxiti+1 − ti < δ it holds |S (f, µ) − S | < ε.The existence of the Stieltjes integral is guaranteed, if f is continuous and µ is of boundedvariation on [a, b]. Notice, that the Riemann integral is a special case of the Stieltjes integral forµ(t) = t. Some properties of the Stieltjes integral are summarized below, cf. Natanson [Nat75]and Widder [Wid46].

• If one of the integrals b

a f (t)dµ(t) or b

a µ(t)df (t) exists, then so does the other and itholds b

af (t)dµ(t) +

b

aµ(t)df (t) = [f (t)µ(t)]b

a .

This is the partial integration rule .

• It holds b

adµ(t) = µ(b) − µ(a).

If µ is constant then b

af (t)dµ(t) = 0.

• If f is continuous and µ is of bounded variation in [a, b] then

F (t) = t

a

f (τ )dµ(τ ), a≤

t ≤

b

is of bounded variation. Moreover

F (t+) − F (t) = f (t)(µ(t+) − µ(t)), a ≤ t < b,

F (t) − F (t−) = f (t)(µ(t) − µ(t−)), a < t ≤ b.

• If f is continuous, g ∈ L1([a, b],R) and

µ(t) =

t

cg(τ )dτ, a ≤ c ≤ b, a ≤ t ≤ b,

then µ is of bounded variation on (a, b) and b

af (t)dµ(t) =

b

af (t)g(t)dt =

b

af (t)µ(t)dt.

The latter integral is a Lebesgue integral.

If g is continuous, h is of bounded variation in [a, b] and

µ(t) =

t

cg(τ )dh(τ ), a ≤ c ≤ b, a ≤ t ≤ b,

then

b

af (t)dµ(t) =

b

af (t)g(t)dh(t).

c 2006 by M. Gerdts

Page 31: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 31/243

2.4. STIELTJES INTEGRAL 23

• If f is of bounded variation and µ is absolutely continuous on [a, b], then b

af (t)dµ(t) =

b

af (t)µ(t)dt,

where the integral on the right is a Lebesgue integral.• If f is continuous and µ is monotonically increasing, then there exists a ξ ∈ [a, b] such that b

af (t)dµ(t) = f (ξ )(µ(b) − µ(a)).

This is a Mean-value theorem .

Lemma 2.4.1 Let f be continuous and µ monotonically increasing on [a, b]. Let µ be differen-tiable at t. Then, it holds

d

dt

t

af (s)dµ(s) = f (t)

d

dtµ(t).

Proof. Let µ be differentiable at t. Define

F (t) :=

t

af (s)dµ(s).

The Mean-value theorem yields

F (t + h) − F (t) =

t+h

tf (s)dµ(s) = f (ξ h)(µ(t + h) − µ(t)), t ≤ ξ h ≤ t + h.

Hence,

limh→0

F (t + h) − F (t)

h = lim

h→0f (ξ h) · µ(t + h) − µ(t)

h = f (t) · d

dtµ(t).

Observe, that t ≤ ξ h ≤ t + h → t holds.

2.4.1 Continuous Functions

The space C ([a, b],R) of continuous functions endowed with the norm

f ∞ = maxa≤t≤b

|f (t)|

is a Banach space. We turn our attention to the representation of elements of the dual spaceof C ([a, b],R). Let µ be a function of bounded variation on [a, b]. Then, for every continuousfunction f on [a, b] the number

Φ(f ) := b

a

f (t)dµ(t)

exists. A closer investigation of the mapping Φ : C ([a, b],R) → R reveals, that Φ is a bounded,linear functional (and hence a continuous functional) on the space of continuous functionsC ([a, b],R). Furthermore, all continuous linear functionals on C ([a, b],R) have the represen-tation Φ, cf. Natanson [Nat75], p. 266. It holds

Theorem 2.4.2 (Riesz, cf. Natanson [Nat75], p. 266)Let Φ : C ([a, b],R) → R be a linear, continuous functional. Then there exists a function µ of bounded variation on [a, b] such that for every f ∈ C ([a, b],R) it holds

Φ(f ) = b

af (t)dµ(t).

c 2006 by M. Gerdts

Page 32: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 32/243

24 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

µ is defined a.e. in [a, b] with exception of an additive constant, cf. Gopfert and Riedrich [GR80],p. 59, Luenberger [Lue69], p. 113.The normalized space of functions of bounded variations N BV ([a, b],R) consists of all functionsµ of bounded variation which are continuous from the right on (a, b) and satisfy µ(a) = 0. Anorm is given by

µ

= T V (f ,a,b). With this definition, the correspondence between the dualof C ([a, b],R) and N BV ([a, b],R) is unique.

2.5 Set Arithmetic

Let S, T ⊆ X be sets and λ ∈ R. The arithmetic operations ‘+’, ‘−’, and ‘·’ are defined as

S ± T := s ± t | s ∈ S, t ∈ T , λS := λs | s ∈ S .

If S, T are convex, then S + T and λS are also convex.A set S is called a cone with vertex Θ, if x ∈ S implies αx ∈ S for all α ≥ 0. If S is a cone withvertex Θ, then x0 + S is a cone with vertex at x0. Let S ⊆ X be a set and x ∈ S . The set

cone(S, x) := α(z − x) | α ≥ 0, z ∈ S = R+ · (S −x)

is called conical hull of S − x. If S is a cone with vertex at Θ, the conical hull of S − x canbe written as

cone(S, x) = z − αx | α ≥ 0, z ∈ S = S − αx | α ≥ 0 = S −R+ · x.

If S is a convex cone with vertex at Θ then it holds

cone(S, x) = S + αx | α ∈R = S + R · x.

To show the latter, notice that cone(S, x) ⊆ S +R · x. Hence, it suffices to show that given an

element y ∈ S +R · x it can be written as s − αx with s ∈ S and α ≥ 0. So, let y = s1 + α1x,s1 ∈ S , α1 ∈ R. If α1 ≤ 0, we are ready. Hence, assume α1 > 0. From

y = s1 + α1x + x − x = s1 + (1 + α1)x =:s2∈S

−x = 2

1

2s1 +

1

2s2

∈S

−x

it follows y ∈ cone(S, x).Let S ⊆ X be a set. The positive dual cone of S is defined as

S + := x∗ ∈ X ∗ | x∗(x) ≥ 0 ∀x ∈ S .

The negative dual cone of S is defined as

S − := x∗ ∈ X ∗ | x∗(x) ≤ 0 ∀x ∈ S .

S is called affine set , if (1 − λ)x + λy ∈ S ∀x, y ∈ S, λ ∈ R.

The set

aff(S ) :=

M | M is an affine set and S ⊆ M

= m

i=1

λixi | m ∈ N,m

i=1

λi = 1, xi ∈ S, i = 1, . . . , m .

c 2006 by M. Gerdts

Page 33: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 33/243

2.6. SEPARATION THEOREMS 25

is called affine hull of S . The set

relint(S ) := x | ∃ε > 0 : U ε(x) ∩ aff(S ) ⊆ S

is called relative interior of S . The set cl(S )

\relint(S ) is called relative boundary of S .

The following results for X = Rn can be found in Section 6 of Rockafellar [Roc70]. Let S ⊆ Rn

be convex, x ∈ relint(S ), and y ∈ cl(S ). Then, (1−λ)x +λy ∈ relint(S ) for 0 ≤ λ < 1. If S ⊆ Rn

is convex and S = ∅ then relint(S ) = ∅. Furthermore, the three convex sets relint(S ), S , andcl(S ) have the same affine hull, the same relative interior and the same closure. Furthermore,cl(relint(S )) = cl(S ) and relint(cl(S )) = relint(S ). Let A : Rn → Rm be a linear transformationand S ⊆ Rn convex. Then relint(A(S )) = A(relint(S )) and A(cl(S )) ⊆ cl(A(S )). For any convexsets S, T ⊆ Rn it holds

relint(S + T ) = relint(S ) + relint(T ), cl(S ) + cl(T ) ⊆ cl(S + T ).

2.6 Separation Theorems

The proof of the necessary Fritz-John conditions is based on the separability of two convex setsby a hyperplane.

Let f : X → R be a non-zero, linear functional and γ ∈ R. Then the set

H := x ∈ X | f (x) = γ

is called a hyperplane . If f is continuous, the hyperplane is closed. The set

H + := x ∈ X | f (x) ≥ γ

is called positive half space of H . Similarly,

H − := x ∈ X | f (x) ≤ γ

is the negative half space of H .

A hyperplane separates the sets A and B , if

f (x) ≤ γ, ∀x ∈ A

f (x) ≥ γ, ∀x ∈ B.

A hyperplane strictly separates the sets A and B , if

f (x) < γ, ∀x ∈ Af (x) > γ, ∀x ∈ B.

A hyperplane H properly separates the sets A and B , if not both sets are contained in H .

Theorem 2.6.1 Let T : X → Y be a linear and continuous operator and R := im (T ) ⊂ Y the image of T . Furthermore, assume that y0 ∈ Y is a point with y0 ∈ R. Then there exists a linear

functional f with f (r) = 0 for r ∈ R and f (y0) = 0.

Remark 2.6.2 Notice, that f is not necessarily continuous. The hyperplane y | f (y) = 0separates R and y0.

c 2006 by M. Gerdts

Page 34: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 34/243

26 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

Proof. Consider the linear subspace L := ty0 | t ∈ R ⊂ Y . It holds L = ∅ and L = Y , sinceL ⊂ R. Hence, Y can be written as Y = L ⊕ S with a subspace S ⊂ Y . Define the functionalf : Y → R by

f (s + ty0) := t.

Then, f (s) = 0 for s ∈ S and f (y0) = 1. Furthermore, f is linear because

f (λ(s + ty0)) = f ( λs ∈S

+ λt ∈R

y0) = λt = λf (s + ty0),

f ((s1 + t1y0) + (s2 + t2y0)) = f (s1 + s2 ∈S

+ (t1 + t2) ∈R

y0) = t1 + t2

= f (s1 + t1y0) + f (s2 + t2y0).

This completes the proof.

If the subspace is closed, every point not contained in this subspace can be separated by a

continuous non-zero linear functional.

Theorem 2.6.3 Let X be a Banach space, M a closed subspace of X , and x ∈ X , x ∈ M .Then there exists f ∈ X ∗ with f (x) = 0 and f (x) = 0 for all x ∈ M .

Proof. cf. Werner [Wer95], Cor. III.1.8, p.98.

The following results are concerned with separation in Rn and can be found in Section 11 of Rockafellar [Roc70].

Theorem 2.6.4 Let A, B ⊆ Rn be non-empty convex sets. There exists a hyperplane separating A and B properly, if and only if relint (A)

∩relint (B) =

∅.

Theorem 2.6.5 Let A, B ⊆ Rn be non-empty convex sets. There exists a hyperplane separating A and B strictly, if and only if Θn ∈ cl (A − B), i.e. if

inf a − b | a ∈ A, b ∈ B > 0.

We are in order to list separation theorems for vector spaces, which can be found in Lem-pio [Lem71b, Lem71a]. Notice, that relint and int are to be understood in the purely algebraicsense and that the functional defining the hyperplane is not necessarily continuous.

Theorem 2.6.6 Let A and B be non-empty convex subsets of a vector space X . If relint (A) = ∅,A has finite defect, and relint (A)

∩ B =

∅, then there exists a non-zero linear functional f

separating A and B.

Theorem 2.6.7 Let A and B be non-empty convex subsets of a vector space X with int (A) = ∅.Then there exists a non-zero linear functional f separating A and B if and only if int (A)∩B = ∅.

Theorem 2.6.8 Let A and B be non-empty convex subsets of a vector space X with relint (A) =∅ and relint (B) = ∅. Then there exists a non-zero linear functional f separating A and B if and only if either A ∪ B are contained in one hyperplane or relint (A) ∩ relint (B) = ∅.

The subsequent separation theorems in Banach spaces can be found in Bonnans and Shapiro [BS00].Here, the functional defining the hyperplane is continuous and hence an element of the dualspace.

c 2006 by M. Gerdts

Page 35: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 35/243

2.7. DERIVATIVES 27

Theorem 2.6.9 (cf. Bonnans and Shapiro [BS00], Th. 2.13)Let A and B be convex subsets of a Banach space X , where A has non-empty interior. Then there exists a non-zero functional f ∈ X ∗ separating A and B if and only if int (A) ∩ B = ∅.

Theorem 2.6.10 (cf. Bonnans and Shapiro [BS00], Th. 2.14)

Let A and B be closed convex subsets of a Banach space X , where A is compact and A ∩ B = ∅.Then there exists a non-zero functional f ∈ X ∗ separating A and B strictly.

Theorem 2.6.11 (cf. Bonnans and Shapiro [BS00], Th. 2.17)Let A be a convex subset of a Banach space X with non-empty relative interior and x0 ∈ X a point with x0 ∈ relint (A). Then there exists a non-zero functional f ∈ X ∗ separating A and x0.

The proofs exploit the Hahn-Banach theorem (cf. Werner [Wer95], Th. III.1.2, Luenberger [Lue69],p. 111).

Theorem 2.6.12 (Hahn-Banach)Let X be a vector space, M a linear subspace of X , f : M → R a linear functional on M , and g : X → R a subadditive and positively homogeneous function, i.e. a function satisfying

g(x + y) ≤ g(x) + g(y),

g(αx) = αg(x),

for all α ≥ 0 and x, y ∈ X . Furthermore, let g majorize f on M , i.e.

f (x) ≤ g(x) ∀x ∈ M.

Then there exists a linear functional f : X → R with

f (x) = f (x), ∀

x∈

M,

f (x) ≤ g(x), ∀x ∈ X.

As a consequence of the Hahn-Banach theorem it follows that continuous linear functionalsdefined on a linear subspace of X can be extended to the whole space X . This is becausecontinuous linear functionals are bounded, i.e. it holds |f (x)| ≤ C xX on M . The subadditiveand positively homogeneous function C ·X plays the role of the function g in the Hahn-Banachtheorem.

2.7 Derivatives

Definition 2.7.1

(i) F : X → Y is called directionally differentiable at x in direction h ∈ X if the limit

F (x; h) = limt↓0

1

t (F (x + th) − F (x)))

exists. F (x; h) is called directional derivative of F at x in direction h.

(ii) F : X → Y is called Gateaux-differentiable at x, if there exists a continuous and linear operator δF (x) : X → Y with

limt↓0

F (x + th) − F (x) − tδF (x)(h)

t = ΘY

for all h ∈

X . The operator δF (x) is called Gateaux differential of F at x.

c 2006 by M. Gerdts

Page 36: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 36/243

28 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

(iii) F : X → Y is called Frechet-differentiable at x ∈ X or differentiable at x ∈ X , if there exists a continuous linear operator F (x) : X → Y with

limhX→0

F (x + h) − F (x) − F (x)(h)

h

X

= ΘY

for all h ∈ X .

F is called regular at x ∈ X , if F is Frechet-differentiable and im (F (x)) = Y .

If x → F (x) is continuous in the strong topology of L(X, Y ), then F is called continuouslydifferentiable.

(iv) F : X × Y → Z is called (partially) Frechet-differentiable w.r.t. x at (x∗, y∗) ∈ X × Y , if F (·, y∗) is Frechet-differentiable at x∗. We denote the partial derivative of F w.r.t. x at (x∗, y∗) by F x(x∗, y∗). A similar definition holds for the component y.

If F : X × Y → Z is Frechet-differentiable at (x∗, y∗), then F is also (partially) Frechet-differentiable w.r.t. x and y and it holds

F (x∗, y∗)(x, y) = F x(x∗, y∗)(x) + F y(x∗, y∗)(y)

for all (x, y) ∈ X × Y .The following statements can be found in Ioffe and Tihomirov [IT79].

• If F : X → Y is Gateaux-differentiable at x, then F (x; h) = δF (x)(h).

• If F : X → Y is Frechet-differentiable at x, then F is continuous and Gateaux-differentiableat x with F (x)(h) = δF (x)(h).

• Let F : X → Y be continuous in a neighborhood of x0 ∈ X and let F be Gateaux-differentiable for every x in this neighborhood. Furthermore, let the mapping x → δF (x)

be continuous. Then F is Frechet-differentiable in this neighborhood and F

(x) = δF (x).

• Chain rule for H = G F : It holds H (x) = G(F (x)) F (x).

• Mean-value theorem (cf. Ioffe and Tihomirov [IT79], p. 27):

Let X , Y be linear topological spaces, F Gateaux-differentiable for every point in [x, x +h] ⊆ U and z → F (z)(h) continuous. Then

F (x + h) − F (x) =

10

F (x + th)(h)dt.

If X , Y are Banach spaces, it holds

F (x + h) − F (x) ≤ sup0≤t≤1 F

(x + th) · h.

• Implicit function theorem (cf. Ioffe and Tihomirov [IT79], p. 29):

Let X , Y , Z be Banach spaces and F : U ⊆ X × Y → Z continuously differentiable. LetF (x0, y0) = ΘZ and let the partial derivative F y(x0, y0) be regular. Then there exists someneighborhood V of x0 and a mapping y : V → Y with y (x0) = y0 and

F (x, y(x)) = ΘZ for all x ∈ V.

Moreover, y(·) is Frechet-differentiable in V and

y(x) =

−(F y(x, y(x)))−1(F x(x, y(x))) for all x

∈ V.

c 2006 by M. Gerdts

Page 37: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 37/243

2.8. VARIATIONAL EQUALITIES AND INEQUALITIES 29

2.8 Variational Equalities and Inequalities

We summarize some results on variational equalities and inequalities. These results are exploitedduring the proof of the local minimum principles for optimal control problems.

Lemma 2.8.1 Let f, g ∈

L∞([t0, tf ],R), s ∈

C ([t0, tf ],R), and µ∈

BV ([t0, tf ],R). If tf

t0

f (t)h(t) + g(t)h(t)dt +

tf

t0

s(t)h(t)dµ(t) = 0

holds for every h ∈ W 1,∞([t0, tf ],R) with h(t0) = h(tf ) = 0, then there exists a function

g ∈ BV ([t0, tf ],R) such that g(t) = g(t) a.e. in [t0, tf ] and g(t) = C + t

t0f (τ )dτ +

tt0

s(τ )dµ(τ ).

Proof. Define F (t) := t

t0f (τ )dτ . Then tf

t0

f (t)h(t)dt =

tf

t0

h(t)dF (t) = [F (t)h(t)]tf t0 −

tf

t0

F (t)dh(t) = − tf

t0

F (t)h(t)dt,

since h(t0) = h(tf ) = 0. Similarly, with G(t) := tt0 s(τ )dµ(τ ) we find tf

t0

s(t)h(t)dµ(t) =

tf

t0

h(t)dG(t) = [G(t)h(t)]tf t0 −

tf

t0

G(t)dh(t) = − tf

t0

G(t)h(t)dt.

Furthermore, for any constant c and every function h ∈ W 1,∞([t0, tf ],R) with h(t0) = h(tf ) = 0,it holds tf

t0

ch(t)dt = 0.

Hence, we conclude that

0 = tf

t0 f (t)h(t) + (g(t) + c)h(t) dt + tf

t0

s(t)h(t)dµ(t)

=

tf

t0

(−F (t) − G(t) + g(t) + c) h(t)dt

holds for every h ∈ W 1,∞([t0, tf ],R) with h(t0) = h(tf ) = 0. Choose

c = 1

tf − t0

tf

t0

(F (t) + G(t) − g(t)) dt,

h(t) =

t

t0

(−F (τ ) − G(τ ) + g(τ ) + c) dτ.

Observe, that h(t0

) = h(tf

) = 0. Then,

0 =

tf

t0

f (t)h(t) + g(t)h(t)

dt +

tf

t0

s(t)h(t)dµ(t)

=

tf

t0

(−F (t) − G(t) + g(t) + c) h(t)dt

=

tf

t0

(−F (t) − G(t) + g(t) + c)2 dt.

This implies −F (t) − G(t) + g(t) + c = 0 a.e. in [t0, tf ] and thus g(t) = F (t) + G(t) − c =

t

t0f (τ )dτ +

t

t0s(τ )dµ(τ ) − c a.e. in [t0, tf ]. Setting g(t) = F (t) + G(t) − c yields the assertion.

c 2006 by M. Gerdts

Page 38: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 38/243

30 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

Remark 2.8.2 Replacing F,G, h, and c in the proof by tf

t f (τ )dτ , tf

t s(τ )dµ(τ ), tf

t (F (τ ) +

G(τ ) + g(τ ) + c)dτ , and 1tf −t0

tf t0

(−F (t) − G(t) − g(t))dt, respectively, yields g(t) = −F (t) −G(t) − c = − tf

t f (τ )dτ − tf t s(τ )dµ(τ ) − c.

Setting g(t) ≡ 0 and µ ≡ 0 in the previous result and exploiting W 1,∞([t0, tf ],R) ⊂ L∞([t0, tf ],R),we immediately obtain the following special result.

Lemma 2.8.3 Let f ∈ L∞([t0, tf ],R). If tf

t0

f (t)h(t)dt = 0

holds for every h ∈ L∞([t0, tf ],R), then f (t) = 0 a.e. in [t0, tf ].

Lemma 2.8.4 Let f ∈ L1([t0, tf ],R). If

tf

t0f (t)h(t)dt ≥ 0

holds for every h ∈ L∞([t0, tf ],R) with h(t) ≥ 0 a.e. in [t0, tf ], then f (t) ≥ 0 a.e. in [t0, tf ].

Proof. Assume the contrary, i.e. there is a set M ⊆ [t0, tf ] with measure greater than 0 andf (t) < 0 for t ∈ M . Choose h to be one on M and zero otherwise. Then tf

t0

f (t)h(t)dt =

M

f (t)dt < 0

in contradiction to the assumption.

Lemma 2.8.5 Let µ

∈ BV ([t0, tf ],R). If tf

t0

f (t)dµ(t) ≥ 0

holds for every f ∈ C ([t0, tf ],R), then µ is non-decreasing in [t0, tf ].

Proof. According to the assumption of the Lemma, for every continuous function f it holdsS =

tf t0

f (t)dµ(t) ≥ 0. Using the definition of the Riemann-Stieltjes integral, this is equivalentwith the following. For every continuous function and every ε > 0 there exists δ > 0 such thatfor every partition GN = t0 < t1 < .. . < tN = tf with maxi=1,...,N ti − ti−1 < δ it holds

−ε

≤ −ε + S <

N

i=1

f (ξ i) (µ(ti)

−µ(ti−1)) < ε + S, (2.8.1)

where ξ i is an arbitrary point in [ti−1, ti].Now, assume that µ is not non-decreasing. Then there are points t < t with µ(t) = µ(t) + γ ,γ > 0. Let ε := γ /2 and δ > 0 be arbitrary and GN = t0 < t1 < . . . < t p := t < . . . < tq :=t < . . . < tN = tf with maxi=1,...,N ti − ti−1 < δ . Then there exists a continuous non-negativefunction f with f (t) ≡ 1 in [t, t], f (ti) = 0 for i ∈ p, . . . , q , and f linear in [t p−1, t p] and[tq, tq+1]. Then,

N i=1

f (ti−1) (µ(ti) − µ(ti−1)) = µ(t) − µ(t) = −γ < −ε.

This contradicts (2.8.1).

c 2006 by M. Gerdts

Page 39: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 39/243

2.8. VARIATIONAL EQUALITIES AND INEQUALITIES 31

Lemma 2.8.6 Let µ ∈ BV ([t0, tf ],R) be non-decreasing and f ∈ C ([t0, tf ],R) non-positive. If tf

t0

f (t)dµ(t) = 0

holds, then µ is constant on every interval [a, b] ⊆ [t0, tf ] with a < b and f (t) < 0 in [a, b].

Proof. Since S = 0 by assumption we find that for every ε > 0 there exists δ > 0 such thatfor every partition GN = t0 < t1 < .. . < tN = tf with maxi=1,...,N ti − ti−1 < δ it holds

−ε <N

i=1

f (ξ i) (µ(ti) − µ(ti−1)) < ε, (2.8.2)

where ξ i is an arbitrary point in [ti−1, ti].Assume, that µ is not constant in intervals [a, b] of the above type. Then, there exist points t < twith f (t) ≤ −α < 0 for all t ∈ [t, t] and µ(t) = µ(t) − γ , γ > 0 (because µ is non-decreasing).

Choose ε := αγ/2. Let δ > 0 be arbitrary and GN = t0 < t1 < . . . < t p = t < . . . < tq = t <.. . < tN = tf with maxi=1,...,N ti − ti−1 < δ . Then,

N i=1

f (ξ i−1) (µ(ti) − µ(ti−1)) =

pi=1

f (ξ i−1) ≤0

(µ(ti) − µ(ti−1)) ≥0

+

qi= p+1

f (ξ i−1) (µ(ti) − µ(ti−1))

+N

i=q+1

f (ξ i−1)

≤0

(µ(ti) − µ(ti−1))

≥0

≤q

i= p+1

f (ξ i−1) (µ(ti) − µ(ti−1))

≤ −α

qi= p+1

(µ(ti) − µ(ti−1))

= −α

µ(t) − µ(t)

= −αγ < −ε.

This contradicts (2.8.2).

c 2006 by M. Gerdts

Page 40: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 40/243

32 CHAPTER 2. BASICS FROM FUNCTIONAL ANALYSIS

c 2006 by M. Gerdts

Page 41: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 41/243

Page 42: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 42/243

34 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Problem 3.1.2 (Standard Nonlinear Optimization Problem)Let f : X → R be a functional, g : X → Y , h : X → Z operators, S ⊆ X a closed, convex set,and K ⊆ Y a closed convex cone with vertex at Θ Y . Find x such that f (x) is minimized subject to the constraints x ∈ S , g(x) ∈ K , and h(x) = Θ Z .

Notice, that the admissible set in Problem 3.1.2 is given by

Σ = S ∩ g−1(K ) ∩ h−1(ΘZ ),

where g−1(K ) := x ∈ X | g(x) ∈ K denotes the preimage of K under g andh−1(ΘZ ) := x ∈ X | h(x) = ΘZ is the preimage of ΘZ under h.In order to be able to derive necessary conditions for a local minimum of Problem 3.1.2 the setS ⊆ X usually cannot be an arbitrary set but has to fulfill additional conditions, e.g. S has tobe closed and convex with nonempty interior. Nevertheless, there are also practically importantproblems, e.g. optimal control problems or mixed integer optimization problems, where S is adiscrete set. Fortunately, it is possible to derive necessary conditions for such problems as well,but among other things additional convexity or differentiability conditions have to be imposed on

the remaining constraints and the objective function. Application of these necessary conditionsto optimal control problems will result in the famous maximum principle for optimal controlproblems.

3.2 Existence of a Solution

The existence of a solution for the optimization problem 3.1.1 where Σ ⊆ X is a compact setand f is lower semicontinuous, is ensured by the famous Weierstrass Theorem:

Theorem 3.2.1 (Weierstrass)Let Σ be a compact subset of a normed vector space X and let f : Σ → R be lower semicontinuous.Then, f achieves its minimum on Σ.

Proof. (cf. Luenberger [Lue69], p. 40)Assume, that f is not bounded from below on Σ. Then, there is a sequence xi in Σ withf (xi) ≤ −i. Since Σ is compact, there exists a convergent subsequence limk→∞ xik = x withf (xik) ≤ −ik for all k ∈ N. Since f is lower semicontinuous it follows f (x) ≤ lim inf k→∞ f (xik).Hence, f (xik) is bounded from below by f (x) ∈ R. This contradicts f (xik) ≤ −ik → −∞. Thisshows that f is bounded from below on Σ.This in turn implies that f = inf x∈Σ f (x) is a real number and for any i ∈ N there exists axi ∈ Σ with f (xi) ≤ f + 1

i . Since Σ is compact, there exists a convergent subsequence x ik → x

with f (xik) ≤ f + 1ik

for all k ∈ N. Since f is lower semicontinuous it follows f ≤ f (x) ≤lim inf k→∞ f (xik) ≤ f . Hence, f assumes its minimum on Σ.

A generalization is given by, cf. Alt [Alt02],

Theorem 3.2.2

Let Σ ⊆ X and let f : Σ → R be a lower semicontinuous function on Σ. Let the set

lev(f, f (w)) ∩ Σ = x ∈ Σ | f (x) ≤ f (w)be nonempty and compact for some w ∈ Σ. Then, f achieves its minimum on Σ.

Proof. According to the Theorem of Weierstrass there exists x ∈ lev(f, f (w)) ∩ Σ withf (x) ≤ f (x) for all x ∈ lev(f, f (w)) ∩ Σ. For x ∈ Σ\(lev(f, f (w)) ∩ Σ) = Σ\lev(f, f (w)) it holdsf (x) > f (w)

≥ f (x). Hence, x is a minimum of f on Σ.

c 2006 by M. Gerdts

Page 43: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 43/243

3.3. CONICAL APPROXIMATION OF SETS 35

3.3 Conical Approximation of Sets

Conical approximations to sets play an important role in the formulation of necessary conditionsfor constrained optimization problems. We will summarize some important cones. Throughoutthis section, Σ denotes a non-empty set in X .

The (sequential) tangent cone to Σ at x ∈ Σ is given by

T (Σ, x) =

d ∈ X

there exist sequences αkk∈N, αk ↓ 0 andxkk∈N, xk ∈ Σ with limk→∞ xk = x, such thatlimk→∞(xk − x)/αk = d holds.

. (3.3.1)

Figure 3.1 illustrates the tangent cone in several situations.

Figure 3.1: Tangent cones to different sets.

The tangent cone is a closed cone with vertex at zero. If x happens to be an interior point of Σ,then T (Σ, x) = X . If Σ is a convex set, then the tangent cone is convex and can be written as

T (Σ, x) = d ∈ X | ∃αk ↓ 0, dk → d : x + αkdk ∈ Σ= d ∈ X | ∃α > 0 : x + αd ∈ Σ.

It is easy to show, cf. Kirsch et al. [KWW78], p. 31, that the tangent cone can be written as

T (Σ, x) =

d ∈ X

there exist σ > 0 and a mapping r : (0, σ] → X with

limε↓0r(ε)

ε = ΘX and a sequence αkk∈N, αk ↓ 0 withx + αkd + r(αk) ∈ Σ for all k ∈ N.

.

Theorem 3.3.1 (Ljusternik)

Let X, Z be Banach spaces, h : X → Z a mapping, and x ∈ M := x ∈ X | h(x) = ΘZ . Let

h be continuous in a neighborhood of x and continuously Gateaux-differentiable at x. Let the Gateaux-derivative δh(x) be surjective. Let d ∈ X be given with δh(x)(d) = ΘZ . Then, there exist ε0 > 0 and a mapping

r : (0, ε0] → X, limε↓0

r(ε)

ε = ΘX

such that h(x + εd + r(ε)) = ΘZ

holds for every ε ∈ (0, ε0]. In particular, it holds

d

∈ X

|δh(x)(d) = ΘZ

= T (M, x).

c 2006 by M. Gerdts

Page 44: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 44/243

36 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Proof. cf. Kirsch et al. [KWW78], p. 40

Recall, that given a cone C ⊆ X with vertex at zero, the positive and negative dual cones aregiven by

C + :=

x∗

∈ X ∗

|x∗(x)

≥ 0 for all x

∈ C

,

C − := x∗ ∈ X ∗ | x∗(x) ≤ 0 for all x ∈ C ,

respectively. Sometimes these cones are also called positive and negative polar cones or normalcones or conjugate cones, respectively. Functionals from C + are called positive on C , andfunctionals from C − are called negative on C . C + and C − are non-empty closed convex cones.

Σ

T (Σ, x)−

T (Σ, x)

x

Figure 3.2: Tangent cone to Σ and its negative dual cone in X = Rn.

The upcoming necessary conditions involve a conic approximation of the set S and a linearizationof the constraint g(x) ∈ K in Problem 3.1.2.Let g be Frechet-differentiable. The linearizing cone of K and S at x is given by

T lin(K,S,x) := d ∈ cone(S, x) | g(x)(d) ∈ cone(K, g(x)), h(x)(d) = ΘZ .

A relation between tangent cone and linearizing cone is given by

Corollary 3.3.2 Let g and h be Frechet-differentiable. Let cone (S, x) and cone (K, g(x)) be closed. Then it holds T (Σ, x) ⊆ T lin(K,S,x).

Proof. Let d ∈ T (Σ, x). Then, there are sequences εk ↓ 0 and xk → x with xk ∈ S, g(xk) ∈K, h(xk) = ΘZ and (xk − x)/εk → d. Since (xk − x)/εk ∈ cone(S, x) and cone(S, x) is closed itholds d ∈ cone(S, x). The continuity of h (x)(·) implies

ΘZ = limk→∞

h(x)((xk − x)/εk) = h(x)(d).

Furthermore, since g is Frechet-differentiable it holds

g(xk) = g(x) + g(x)(xk

−x) + αk

xk

−x

X

c 2006 by M. Gerdts

Page 45: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 45/243

3.4. FIRST ORDER NECESSARY CONDITIONS OF FRITZ-JOHN TYPE 37

with some sequence αkk∈N ⊆ Y , αk → ΘY . Hence,

g(xk) − g(x)

εk

∈cone(K,g(x))

= g (x)

xk − x

εk

+ αk

xk − x

εk

X

.

Since cone(K, g(x)) is closed, the term on the left converges to an element in cone(K, g(x)),while the term on the right converges to g (x)(d). Thus, g(x)(d) ∈ cone(K, g(x)). Togetherwith d ∈ cone(S, x) and h (x)(d) = ΘZ this shows d ∈ T lin(K,S,x).

3.4 First Order Necessary Conditions of Fritz-John Type

First, we will discuss a geometrically motivated first order necessary condition for Problem 3.1.1,which involves the tangent cone. It holds

Theorem 3.4.1 Let f : X → R be Frechet-differentiable at x and let x be a local minimum of Problem 3.1.1. Then

f (x)(d) ≥ 0 ∀d ∈ T (Σ, x).

Proof. Let d ∈ T (Σ, x). Then there are sequences αk ↓ 0, xk → x, xk ∈ Σ with d =limk→∞(xk − x)/αk. Since x is a local minimum and f is Frechet-differentiable at x it follows

0 ≤ f (xk) − f (x) = f (x)(xk − x) + o(xk − xX ).

Division by αk > 0 yields

0 ≤ f (x)

xk − x

αk

→d

+

xk − x

αk

X →dX

· o(xk − xX )

xk − xX

→0

.

The subsequent theorem stating necessary conditions for Problem 3.1.2 provides the basis forthe minimum principle for optimal control problems and can be found in Lempio [Lem72].

Theorem 3.4.2 (First Order Necessary Conditions)

Let f : X → R and g : X → Y be Frechet-differentiable and h : X → Z continuously Frechet-differentiable. Let x be a local minimum of Problem 3.1.2, int (S ) = ∅, and int (K ) = ∅. As-sume that im (h(x)) is not a proper dense subset of Z . Then there exist nontrivial multipliers (l0, λ∗, µ∗) ∈ R× Y ∗ × Z ∗, (l0, λ∗, µ∗) = (0, ΘY ∗ , ΘZ ∗) such that

l0 ≥ 0, (3.4.1)

λ∗ ∈ K +, (3.4.2)

λ∗(g(x)) = 0, (3.4.3)

l0f (x)(d) − λ∗(g(x)(d)) − µ∗(h(x)(d)) ≥ 0, ∀d ∈ S − x. (3.4.4)

Proof. Consider the linearized problem

Minimize f (x) + f (x)(x − x)w.r.t. x ∈ S subject to g(x) + g(x)(x − x) ∈ K,

h(x) + h(x)(x−

x) = ΘZ

.

c 2006 by M. Gerdts

Page 46: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 46/243

38 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

For the first part of the proof we assume that the mapping h(x)(·) : X → Z is surjective andthat there exists a feasible x ∈ int(S ) for the linearized problem satisfying

g(x) + g(x)(x − x) ∈ int(K ), (3.4.5)

h(x) + h(x)(x−

x) = ΘZ

, (3.4.6)

f (x) + f (x)(x − x) < f (x). (3.4.7)

(i) Since h(x) is surjective and h(x)(x − x) = ΘZ holds, we may apply Ljusternik’s theorem.Hence, there exist t0 > 0 and a mapping

r : [0, t0] → X, limt↓0

r(t)

t = ΘX , r(0) := ΘX ,

such thath(x + t(x − x) + r(t)) = ΘZ

holds for every t ∈

[0, t0]. This means, that the curve

x(t) = x + t(x − x) + r(t)

remains feasible for the nonlinear equality constraints for every t ∈ [0, t0]. Furthermore, itholds

x(0) = x, x(0) = x − x.

The latter holds because

x(0) = limt↓0

x(t) − x(0)

t = lim

t↓0

x(t) − x

t = x − x + lim

t↓0

r(t)

t = x − x.

(ii) Now, we consider the inequality constraints at x. Since g is Frechet-differentiable at x itholds

g(x(t)) = g(x) + tg(x)(x − x) + o(t)

= t(g(x) + g(x)(x − x) ∈int(K )

) + ( 1 − t) g(x) ∈K

+o(t).

Since g(x) + g(x)(x − x) ∈ int(K ) there exists δ 1 > 0 such that

U δ1(g(x) + g(x)(x − x)) ⊆ K.

Furthermore, the convexity of K yields

(1 − t)g(x) + ty ∈ K

for all y ∈ U δ1(g(x) + g(x)(x − x)) and all 0 ≤ t ≤ 1. Since for t ↓ 0 it holds o(t)/t → 0there exists δ 2 > 0 with o(t)/tY < δ 1 for all 0 < t < δ 2. Hence,

g(x(t)) = t(g(x) + g(x)(x − x) + o(t)/t ∈U δ1(g(x)+g(x)(x−x))

) + ( 1 − t)g(x) ∈ K.

Hence, for sufficiently small t > 0 the curve x(t) stays feasible for the nonlinear inequalityconstraints.

c 2006 by M. Gerdts

Page 47: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 47/243

3.4. FIRST ORDER NECESSARY CONDITIONS OF FRITZ-JOHN TYPE 39

(iii) Now, the objective function is investigated. It holds

d

dtf (x(t))

t=0= f (x) · dx

dt(0) = f (x)(x − x)

(3.4.7)< 0.

Hence, x − x is a direction of descent of f , i.e. it holds f (x(t)) < f (x) for t > 0 sufficientlysmall. (this will contradict the local minimality of x for Problem 3.1.2).

(iv) The point x in the linear problem fulfilling (3.4.5)-(3.4.7) is assumed to be an interiorpoint of S , cf. Figure 3.3.

Then, since S is assumed to be convex, there exists a neighborhood U δ(x) such thatx + t(z − x) ∈ S for all 0 ≤ t ≤ 1 and all z ∈ U δ(x). Since

limt↓0

r(t)

t = ΘX

there exists ε > 0 with r(t)

t

X

< δ

for all 0 < t < ε. Hence,

x(t) = x + t(x − x) + r(t) = x + t

x + r(t)

t ∈U δ(x)

−x

∈ S

for 0 < t < ε.

Items (i)-(iv) showed the following: If x is a local minimum of Problem 3.1.2 and h

(x) issurjective, then

f (x)(d) ≥ 0 ∀d ∈ T (x) :=

d ∈ int(cone(S, x))

g(x)(d) ∈ int(cone(K, g(x))),h(x)(d) = ΘZ

. (3.4.8)

Consider the non-empty convex set

A :=

f (x)(d) + r

g(x)(d) − kh(x)(d)

d ∈ int(cone(S, x)), k ∈ int(cone(K, g(x))), r > 0

.

If h (x) is not surjective and im(h(x)) is not a proper dense subset of Z , the set

M := cl

r

yh(x)(x)

r ∈ R, y ∈ Y, x ∈ X

is a proper closed subspace of R × Y × Z . According to Theorem 2.6.3 there is a non-zerofunctional λ ∈ (R× Y × Z )∗ with λ(r,y,z) = 0 for all (r,y,z) ∈ M . Hence, the hyperplane

(r,y,z) ∈ R× Y × Z | λ(r,y,z) = 0

trivially separates the sets A and the point (0, ΘY , ΘZ ), since both are contained in M .

c 2006 by M. Gerdts

Page 48: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 48/243

40 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

A can be decomposed into A = A1 + A2 with

A1 :=

f (x)(d)g(x)(d)h(x)(d)

d ∈ int(cone(S, x))

,

A2 :=

r

−kΘZ

k ∈ int(cone(K, g(x))), r > 0

.

If h (x) is surjective, the projection of A1 onto Z contains interior points in Z according to theopen mapping theorem. Hence, A1+A2 contains interior points in R×Y ×Z . The considerationsin (i)-(iv) showed that (0, ΘY , ΘZ ) ∈ int(A). For, let us assume (0, ΘY , ΘZ ) ∈ int(A). Then thereexists d ∈ int(cone(S, x)) with f (x)(d) < 0, g(x)(d) ∈ int(cone(K, g(x))), and h(x)(d) = ΘZ

contradicting (3.4.8).According to the separation theorem 2.6.11 there exists a hyperplane separating A and (0, ΘY , ΘZ ),i.e. there exist multipliers

(0, ΘY ∗ , ΘZ ∗) = (l0, λ∗, µ∗) ∈ R× Y ∗ × Z ∗ = (R× Y × Z )∗

such that

0 ≤ l0

f (x)(d) + r− λ∗

g(x)(d) − y

− µ∗

h(x)(d)

for all d ∈ int(cone(S, x)), y ∈ int(cone(K, g(x))), r > 0. Owing to the continuity of thefunctionals l0(·), λ∗(·), µ∗(·) and the linear operators f (x)(·), g (x)(·), and h(x)(·) this inequalityalso holds for all d ∈ cone(S, x), y ∈ cone(K, g(x)), r ≥ 0.Choosing d = ΘX ∈ cone(S, x) and y = ΘY ∈ cone(K, g(x)) yields l0r ≥ 0 for all r ≥ 0 and

thus l0 ≥ 0. The choices d = ΘX and r = 0 imply λ∗

(y) ≥ 0 for all y ∈ cone(K, g(x)) =k − αg(x) | k ∈ K, α ≥ 0, i.e. λ∗(k − αg(x)) ≥ 0 for all k ∈ K , α ≥ 0. This in turn impliesλ∗ ∈ K + (choose α = 0) and λ∗(g(x)) = 0 (choose k = ΘY , α = 1 and observe that g(x) ∈ K ).

S

x

x

x(t)

Figure 3.3: Fritz-John conditions under set constraints x ∈ S . The assumption int(S ) = ∅ isessential. For sufficiently small t > 0 the curve x(t) stays feasible.

Every point (x, l0, λ∗, µ∗) ∈ X ×R×Y ∗×Z ∗, (l0, λ∗, µ∗) = Θ satisfying the Fritz-John conditions(3.4.1)-(3.4.4) is called Fritz-John point of Problem 3.1.2. The multipliers l0, λ∗, and µ∗ are

c 2006 by M. Gerdts

Page 49: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 49/243

3.5. CONSTRAINT QUALIFICATIONS 41

called Lagrange multipliers or simply multipliers . The main statement of the theorem is thatthere exists a nontrivial vector (l0, λ∗, µ∗) = Θ. Notice, that (l0, λ∗, µ∗) = Θ trivially fulfills theFritz-John conditions. Unfortunately, the case l0 = 0 may occur. In this case, the objectivefunction f does not enter into in the Fritz-John conditions. In case of l0 = 0 we call theFritz-John point (x, l

0, λ∗, µ∗) Karush-Kuhn-Tucker (KKT) point .

First order necessary conditions of Fritz-John type for general cone constrained problems

f (x) → min s.t. x ∈ S, g(x) ∈ K

can be found in Kurcyusz [Kur76]. Essentially, the existence of non-trivial multipliers (defininga separating hyperplane) can be guaranteed, if

im(g(x)) + cone(K, g(x))

is not a proper dense subset of Y .

3.5 Constraint Qualifications

Conditions which ensure that the multiplier l0 in Theorem 3.4.2 is not zero are called regu-larity conditions or constraint qualifications . In this case, without loss of generality l0 can benormalized to one due to the linearity in the multipliers.The following regularity condition was postulated by Robinson [Rob76] in the context of stabilityanalysis for generalized inequalities.

Definition 3.5.1 (Regularity condition of Robinson [Rob76])The regularity condition of Robinson is satisfied at x if

ΘY

ΘZ ∈ int

g(x) + g(x)(x − x) − kh(x)(x

− x)

x ∈ S, k ∈ K

. (3.5.1)

The validity of the regularity condition ensures that the multiplier l0 is not zero.

Theorem 3.5.2 (KKT-conditions)Let the assumptions of Theorem 3.4.2 be satisfied. In addition let the regularity condition of Robinson be satisfied at x. Then the assertions of Theorem 3.4.2 hold with l0 = 1.

Proof. Let us assume that the necessary conditions are valid with l0 = 0, i.e.

λ∗(k) ≥ 0, ∀k ∈ K,

λ∗(g(x)) = 0,

λ∗(g(x)(d)) + µ∗(h(x)(d)) ≤

0, ∀

d∈

S −

x

.

These conditions imply

λ∗(g(x) + g(x)(d) − k) + µ∗(h(x)(d)) ≤ 0, ∀d ∈ S − x, k ∈ K.

Hence, the functional η ∗(y, z) := λ∗(y) + µ∗(z) separates (ΘY , ΘZ ) from the set g(x) + g(x)(x − x) − k

h(x)(x − x)

x ∈ S, k ∈ K

since η∗(ΘY , ΘZ ) = 0. But, (ΘY , ΘZ ) is assumed to be an interior point of this set. Hence, aseparation is impossible and the assumption l0 = 0 was wrong.

c 2006 by M. Gerdts

Page 50: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 50/243

42 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Zowe and Kurcyusz [ZK79] show that the regularity condition of Robinson is stable underperturbations of the constraints and ensures the boundedness of the multipliers.A sufficient condition for the regularity condition of Robinson is the surjectivity constraint

qualification.

Corollary 3.5.3 Let x ∈ int (S ) and let the operator

T : X → Y × Z, T := (g(x), h(x))

be surjective. Then the regularity condition of Robinson (3.5.1) holds.

Proof. Since x ∈ int(S ) there exists an open ball U ε(ΘX ) ⊆ S −x. The operator T is linear,continuous, surjective, and T (ΘX ) = (ΘY , ΘZ ). According to the open mapping theorem 2.2.3the set T (U ε(ΘX )) is open in Y × Z . Hence,

ΘY

ΘZ

∈ int

g(x)(x − x)h(x)(x − x)

x ∈ S

.

Since g(x) ∈ K and thus ΘY ∈ K − g(x) it follows the regularity condition of Robinson (3.5.1).

The subsequent Mangasarian-Fromowitz like condition is sufficient for the regularity conditionof Robinson.

Corollary 3.5.4 Let g : X → Y and h : X → Z be Frechet-differentiable at x, K ⊆ Y a closed convex cone with vertex at zero and int(K )

=

∅, g(x)

∈ K , h(x) = ΘZ . Furthermore, let the

following conditions be fulfilled:

(i) Let h(x) be surjective.

(ii) Let there exist some d ∈ int (S −x) with

h(x)(d) = ΘZ , (3.5.2)

g(x)(d) ∈ int (K − g(x)). (3.5.3)

Then the regularity condition of Robinson (3.5.1) holds.

Proof. Since g(x)+g(x)(d)

∈ int(K ) there exists a ball U ε1(g(x)+ g(x)(d)) with radius ε1 > 0

and center g(x) + g (x)(d) which lies in int(K ). Since subtraction viewed as a mapping fromY × Y into Y is continuous and observing, that

g(x) + g(x)(d) = g(x) + g(x)(d) − ΘY ,

there exist balls U ε2(g(x) + g(x)(d)) and U ε3(ΘY ) with

U ε2(g(x) + g(x)(d)) − U ε3(ΘY ) ⊆ U ε1(g(x) + g(x)(d)).

Since g (x) is a continuous linear mapping, there exists a ball U δ1(d) in X with

g(x) + g(x)(U δ1(d)) ⊆

U ε2(g(x) + g(x)(d)).

c 2006 by M. Gerdts

Page 51: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 51/243

3.6. NECESSARY AND SUFFICIENT CONDITIONS IN FINTE DIMENSIONS 43

Since d ∈ int(S − x), eventually after diminishing δ 1 to δ 2 > 0, we find

g(x) + g(x)(U δ2(d)) ⊆ U ε2(g(x) + g(x)(d))

and

U δ2(ˆd) ⊆ int(S − x).

Since h(x) is continuous and surjective by the open mapping theorem, there exists a ball U ε4(ΘZ )in Z with

U ε4(ΘZ ) ⊆ h(x)(U δ2(d)).

Summarizing, we found the following: For every y ∈ U ε3(ΘY ), z ∈ U ε4(ΘZ ) there exists d ∈U δ2(d), in particular d ∈ int(S − x), with

h(x)(d) = z, g(x) + g(x)(d) ∈ U ε2(g(x) + g(x)(d)),

henceg(x) + g(x)(d) − y ∈ U ε1(g(x) + g(x)(d)) ⊆ int(K ),

i.e. there exists k ∈ int(K ) with

g(x) + g(x)(d) − y = k.

Thus, we provedy = g(x) + g(x)(d) − k, z = h(x)(d)

with some k ∈ int(K ), d ∈ int(S − x).

Remark 3.5.5

• Condition (ii) in Corollary 3.5.4 can be replaced by the fol lowing assumption: Let there exist some d ∈ int (cone (S, x)) with

h(x)(d) = ΘZ ,

g(x)(d) ∈ int (cone (K, g(x))).

• It is possible to show that the above Mangasarian-Fromowitz condition is equivalent toRobinson’s condition for problems of type 3.1.2.

Second order necessary and sufficient conditions for infinite optimization problem are discussedin Maurer and Zowe [MZ79] and Maurer [Mau81]. In Maurer [Mau81] also the so-called twonorm discrepancy is addressed. This problem occurs if the sufficient conditions do not hold forthe norm · X . On the other hand, these conditions may be satisfied for an alternative norm · p on the space X . Unfortunately, the appearing functions usually are not differentiable

anymore in this alternative norm and the proof techniques are more intricate.

3.6 Necessary and Sufficient Conditions in Finte Dimensions

In this section we address the finite dimensional case with X = Rnx, Y = Rng , Z = Rnh,S ⊆ Rnx , and continuously differentiable functions

f : Rnx → R,

g = (g1, . . . , gng) : Rnx → Rng ,

h = (h1, . . . , hnh) : Rnx → R

nh.

We will discuss necessary and sufficient conditions as well as the sequential quadratic program-ming (SQP) method for nonlinear optimization problems of the form

c 2006 by M. Gerdts

Page 52: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 52/243

44 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Problem 3.6.1 (Nonlinear Optimization Problem)Find x ∈ Rnx such that f (x) is minimized subject to the constraints

gi(x) ≤ 0, i = 1, . . . , ng,

h j(x) = 0, j = 1, . . . , nh,x ∈ S.

The feasible set of Problem 3.6.1 is

Σ := x ∈ S | gi(x) ≤ 0, i = 1, . . . , ng, h j (x) = 0, j = 1, . . . , nh.

The set

A(x) := i | gi(x) = 0, 1 ≤ i ≤ ngis called index set of active inequality constraints at x ∈ Σ.The subsequent results are collected from the monographs [BS79], [GMW81], [FM90], [Spe93],

[Man94], [GK99], [Alt02], [GK02].By reformulating the infinite Fritz-John conditions in Theorem 3.4.2 for the finite dimensionalproblem 3.6.1 we find the following finite dimensional version of the Fritz-John conditions.Notice, that the assumption that im(h(x)) be not a proper dense subset of Z = Rnh is triviallyfulfilled in finite dimensions, since then the image of a linear mapping is closed. Furthermore,recall that Rn is a Hilbert space and hence can be identified with its dual space. We will usethe Lagrange function

L(x, l0, λ , µ) := l0f (x) +

ngi=1

λigi(x) +

nh j=1

µ jh j(x).

Theorem 3.6.2 (First Order Necessary Conditions, finite case)Let x be a local minimum of Problem 3.6.1 and S closed and convex with int (S ) = ∅. Then there exist multipliers l0 ≥ 0, λ = (λ1, . . . , λng) ∈ Rng and µ = (µ1, . . . , µnh

) ∈ Rnh not all zerosuch that

Lx(x, l0, λ , µ)(x − x) ≥ 0 for all x ∈ S, (3.6.1)

gi(x) ≤ 0, i = 1, . . . , ng, (3.6.2)

h j (x) = 0, j = 1, . . . , nh, (3.6.3)

λigi(x) = 0, i = 1, . . . , ng, (3.6.4)

λi ≥ 0, i = 1, . . . , ng. (3.6.5)

The constraint qualification of Mangasarian-Fromowitz is defined as in Corollary 3.5.4.

Definition 3.6.3 (Constraint Qualification of Mangasarian-Fromowitz)The constraint qualification of Mangasarian-Fromowitz is satisfied at x, if the following condi-tions are fulfilled:

(a) The derivatives h j (x), j = 1, . . . , nh are linearly independent;

(b) There exists a vector d ∈ int (S − x) with

gi(x)(d) < 0 for i

∈ A(x) and h j (x)(d) = 0 for j = 1, . . . , nh.

c 2006 by M. Gerdts

Page 53: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 53/243

3.6. NECESSARY AND SUFFICIENT CONDITIONS IN FINTE DIMENSIONS 45

Although the following theorem is a consequence of Theorem 3.5.2 and Corollary 3.5.4, we willgive an alternative proof for finite dimensions.

Theorem 3.6.4 (Karush-Kuhn-Tucker (KKT) Conditions I)

Let the assumptions of Theorem 3.6.2 be satisfied and let the constraint qualification of Mangasarian-

Fromowitz be fulfilled at x. Then, the assertions of Theorem 3.6.2 hold with l0 = 1.

Proof. Assume, that the constraint qualification of Mangasarian-Fromowitz holds and that theFritz-John conditions hold at x with l0 = 0, that is, there exist multipliers λ i ≥ 0, i = 1, . . . , ng,µ j, j = 1, . . . , nh not all zero with λigi(x) = 0 for all i = 1, . . . , ng and

ngi=1

λigi(x) +

nh j=1

µ jh j(x)

(d) ≥ 0 for all d ∈ S − x. (3.6.6)

Let d denote the vector in the constraint qualification of Mangasarian-Fromowitz, then we find

ngi=1

λigi(x)(d) = i∈A(x)

λigi(x)(d) ≥ 0.

Since λi ≥ 0 and gi(x)(d) < 0 for i ∈ A(x) this inequality can only be valid if λi = 0 holds forevery i ∈ A(x). Since λi = 0 for i ∈ A(x) we found λi = 0 for all i = 1, . . . , ng.Thus, inequality (3.6.6) reduces to

0 ≤nh

j=1

µ j h j(x)(d) =

nh j=1

µ j(h j (x)(d) − h j(x)(d)) =

nh j=1

µ jh j (x)(d − d)

for all d ∈ S − x. Since d ∈ int(S − x) and h j (x), j = 1, . . . , nh are linearly independent

this inequality can only hold for µ j = 0, j = 1, . . . , nh.Hence, we derived (l0, λ , µ) = 01+ng+nh

, which is a contradiction to the statement that not allmultipliers are zero.

Likewise, the linear independence constraint qualification is defined according to Corollary 3.5.3.Actually, this condition implies that of Mangasarian-Fromowitz.

Definition 3.6.5 (Linear Independence Constraint Qualification (LICQ))The linear independence constraint qualification is satisfied at x, if the following conditions are

fulfilled:

(a) x

∈ int (S );

(b) The derivatives gi(x), i ∈ A(x), and h j (x), j = 1, . . . , nh are linearly independent.

The first part of the following theorem is a consequence of Theorem 3.5.2 and Corollary 3.5.3.Nevertheless, we will give an alternative proof.

Theorem 3.6.6 (Karush-Kuhn-Tucker (KKT) Conditions II)

Let the assumptions of Theorem 3.6.2 be satisfied and let the linear independence constraint qualification be fulfilled at x. Then, the assertions of Theorem 3.6.2 hold with l0 = 1 and in particular

∇xL(x, l0, λ , µ) = 0nx .

Furthermore, the multipliers λ and µ are unique.

c 2006 by M. Gerdts

Page 54: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 54/243

46 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Proof. Again, we assume that the Fritz-John conditions hold with l0 = 0. Since x ∈ int(S ),inequality (3.6.1) has to hold for all x ∈ Rnx and thus

ng

i=1

λigi(x) +

nh

j=1

µ jh j (x) = i∈A(x)

λigi(x) +

nh

j=1

µ jh j(x) = 0nx .

The linear independence of the derivatives implies λi = µ j = 0 for all i = 1, . . . , ng, j = 1, . . . , nh.Again, this is a contradiction to (l0, λ , µ) = 01+ng+nh

.The uniqueness of the Lagrange multipliers follows from the following considerations. Assume,that there are Lagrange multipliers λ i, i = 1, . . . , ng, µ j , j = 1, . . . , nh and λi, i = 1, . . . , ng, µ j,

j = 1, . . . , nh satisfying the KKT conditions. Again, x ∈ int(S ) particularly implies

0nx = f (x) +

ngi=1

λigi(x) +

nh j=1

µ jh j (x),

0nx = f

(x) +

ngi=1

λigi(x) +

nh j=1

µ jh j(x).

Subtracting these equations leads to

0nx =

ngi=1

(λi − λi)gi(x) +

nh j=1

(µ j − µ j )h j(x).

For inactive inequality constraints we have λi = λi = 0, i ∈ A(x). Since the gradients of theactive constraints are assumed to be linearly independent, it follows 0 = λ i − λi, i ∈ A(x), and0 = µ j − µ j, j = 1, . . . , nh. Hence, the Lagrange multipliers are unique.

Remark 3.6.7 There exist several other constraint qualifications. One of the weakest conditions is the constraint qualification of Abadie postulating

T lin(K,S, x) ⊆ T (Σ, x),

where K = y ∈ Rng | y ≤ 0ng and

T lin(K,S,x) = d ∈ cone (S, x) | g i(x)(d) ≤ 0, i ∈ A(x),

h j (x)(d) = 0, j = 1, . . . , nhdenotes the linearizing cone at x.The condition of Abadie is weaker as the previously discussed constraint qualifications since those imply the condition of Abadie but not vice versa.It is important to mention, that the tangent cone T (Σ, x) is independent of the representation of the set Σ by inequality and equality constraints, whereas the linearizing cone T lin(K,S,x)depends on the functions gi and h j describing Σ.

In the sequel, we restrict the discussion to the case S = Rnx . This special case is particularlyconvenient for the design of numerical methods, e.g. SQP methods. Setting S = Rnx , inequality(3.6.1) has to hold for all x ∈ Rnx and thus it is equivalent with the equality

∇xL(x, l0, λ , µ) = 0nx .

The previously discussed regularity conditions additionally ensure l0 = 1.

c 2006 by M. Gerdts

Page 55: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 55/243

3.6. NECESSARY AND SUFFICIENT CONDITIONS IN FINTE DIMENSIONS 47

To decide, whether a given point that fulfills the necessary conditions is optimal we need sufficient conditions . We need the so-called critical cone

T C (x) :=

d ∈ R

nx

gi(x)(d) ≤ 0, i ∈ A(x), λi = 0,gi(x)(d) = 0, i ∈ A(x), λi > 0,

h j (x)(d) = 0, j = 1, . . . , nh

.

The term ‘critical cone‘ is due to the following reasoning. Directions d with g i(x)(d) > 0 forsome i ∈ A(x) or h j (x)(d) = 0 for some j ∈ 1, . . . , nh are infeasible directions. So, consideronly feasible directions d with g i(x)(d) ≤ 0 for i ∈ A(x) and h j(x)(d) = 0 for j = 1, . . . , nh. Forsuch directions the KKT conditions yield

∇f (x)d +

i∈A(x)

λi∇gi(x)d ≤0

+

nh j=1

µ j ∇h j(x)d =0

= 0nx ,

and thus f (x)(d) ≥ 0. If even f (x)(d) > 0 holds, then d is a direction of ascent and thedirection d is not interesting for the investigation of sufficient conditions. So, let f (x)(d) = 0.

This is the critical case. In this critical case it holdsi∈A(x)

λigi(x)(d) =

i∈A(x),λi>0

λigi(x)(d) = 0,

and thus g i(x)(d) = 0 for all i ∈ A(x) with λ i > 0. Hence, d ∈ T C (x) and for such directions weneed additional assumptions about the curvature (2nd derivative!).A second order sufficient condition, cf., e.g., Geiger and Kanzow [GK02], Th. 2.55, p. 67,Alt [Alt02], Th. 7.3.1, p. 281, and Bazaraa et al. [BSS93], Th. 4.4.2, p. 169, is given by

Theorem 3.6.8 (Second Order Sufficient Condition)Let f , gi, i = 1, . . . , ng, and h j , j = 1, . . . , nh be twice continuously differentiable. Let S = Rnx

and let (x, λ, µ) be a KKT point of Problem 3.6.1 with

Lxx(x, λ, µ)(d, d) > 0 ∀d ∈ T C (x), d = 0nx . (3.6.7)

Then there exists a neighborhood U of x and some α > 0 such that

f (x) ≥ f (x) + αx − x2 ∀x ∈ Σ ∩ U.

Proof.

(a) Let d ∈ T (Σ, x), d = 0Rnx . Then there exist sequences x k ∈ Σ, xk → x and αk ↓ 0 with

limk→∞

xk − x

αk= d.

For i ∈ A(x) we have

0 ≥ gi(xk) − gi(x)

αk= gi(ξ k)

xk − x

αk

→ gi(x)(d)

by the mean-value theorem. Similarly, we show h j(x)(d) = 0 for j = 1, . . . , nh. Since

(x, λ, µ) is a KKT point with λi = 0, if gi(x) < 0, we obtain

f (x)(d) = −ng

i=1

λigi(x)(d) −nh

j=1

µ j h j(x)(d) ≥ 0.

Hence, x fulfills the first order necessary condition f (x)(d) ≥

0 for all d ∈

T (Σ, x).

c 2006 by M. Gerdts

Page 56: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 56/243

Page 57: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 57/243

3.7. PERTURBED NONLINEAR OPTIMIZATION PROBLEMS 49

A second order necessary condition involving the critical cone can be found in, e.g., Geiger andKanzow [GK02], Th. 2.54, p. 65.

3.7 Perturbed Nonlinear Optimization ProblemsIn view of the convergence analysis of the SQP method and for real-time approximations we needresults about the sensitivity of solutions under perturbations. We restrict the discussion to thefinite dimensional case. Results for infinite optimization problems can be found in Lempio andMaurer [LM80] and Bonnans and Shapiro [BS00] and the literature cited therein. We investigateparametric optimization problems:

Problem 3.7.1 (Parametric Optimization Problem N LP ( p))Let p ∈ Rnp be a given parameter. Find x ∈ Rnx such that f (x, p) is minimized subject to the constraints

gi(x, p)

≤ 0, i = 1, . . . , ng,

h j(x, p) = 0, j = 1, . . . , nh.

Herein, f, g1, . . . , gng , h1, . . . , hnh : Rnx × Rnp → R are sufficiently smooth functions. Let ˆ p

denote a fixed nominal parameter . We are interested in the behavior of the optimal solutionsx( p) as functions of p in a neighborhood of the nominal parameter ˆ p.The admissible set of N LP ( p) is defined by

Σ( p) := x ∈ Rnx | gi(x, p) ≤ 0, i = 1, . . . , ng, h j(x, p) = 0, j = 1, . . . , nh.

The index set of active inequality constraints is given by

A(x, p) =

i

|gi(x, p) = 0, 1

≤ i

≤ ng

.

Definition 3.7.2 (Strongly Regular Local Solution)

A local minimum x of N LP ( p) is called strongly regular if the following properties hold:

• x is admissible, i.e. x ∈ Σ( p).

• x fulfills the linear independence constraint qualification, i.e. the gradients ∇xgi(x, p),i ∈ A(x, p), ∇xh j(x, p), j = 1, . . . , nh, are linearly independent.

• The KKT conditions hold at (x, λ, µ).

• The strict complementarity condition holds, i.e. λi − gi(x, p) > 0 for all i = 1, . . . , ng.

• The second order sufficient condition (3.6.7) holds.

The following result is based on Fiacco [Fia83, FM90] and Spellucci [Spe93]. Further results canbe found in Bank et al. [BGK+83] and Klatte [Kla90].

Theorem 3.7.3 (Sensitivity Theorem)Let f, g1, . . . , gng , h1, . . . , hnh

: Rnx × Rnp → R be twice continuously differentiable and ˆ p a

nominal parameter. Let x be a strongly regular local minimum of N LP (ˆ p), λ, µ denote the corresponding Lagrange multipliers. Then there exist neighborhoods V (ˆ p) and U δ(x, λ, µ), such that N LP ( p) has a unique strongly regular local minimum

(x( p), λ( p), µ( p)) ∈

U δ(x, λ, µ)

c 2006 by M. Gerdts

Page 58: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 58/243

50 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

for each p ∈ V (ˆ p). Furthermore, it holds A(x, ˆ p) = A(x( p), p). In addition, (x( p), λ( p), µ( p)) is continuously differentiable w.r.t. p with

dx

dp(ˆ p)

dλdp

(ˆ p)

dp(ˆ p)

= −

Lxx (gx) (hx)

Λ · gx Γ Θ

hx Θ Θ

−1

·L

xp

Λ · g p

h p

(3.7.1)

where Λ = diag(λ1, . . . , λng), Γ = diag(g1, . . . , gng ). All functions and their derivatives are

evaluated at (x, λ, µ, ˆ p).

Proof. Consider the nonlinear equation

F (x,λ,µ,p) := L

x(x,λ,µ,p)

Λ

·g(x, p)

h(x, p) = 0nx+ng+nh, (3.7.2)

where Λ := diag(λ1, . . . , λng). F is continuously differentiable and it holds

F (x, λ, µ, ˆ p) = 0nx+ng+nh.

We intend to apply the implicit function theorem. Hence, we have to show the non-singularityof

F x(x, λ, µ, ˆ p) =

Lxx(x, λ, µ, ˆ p) (gx(x, ˆ p)) (hx(x, ˆ p))

Λ · gx(x, ˆ p) Γ Θ

hx(x, ˆ p) Θ Θ

.

In order to show this, we assume without loss of generality, that the index set of active inequalityconstraints is given by A(x, ˆ p) = l+1, . . . , ng, where l denotes the number of inactive inequalityconstraints. Then, the strict complementarity condition implies

Λ =

Θ Θ

Θ Λ2

, Γ =

Γ1 Θ

Θ Θ

,

with non-singular matrices

Λ2 := diag(λl+1, . . . , λng ) and Γ1 := diag(g1(x, ˆ p), . . . , gl(x, ˆ p)).

Consider the linear equation

L

xx(x,ˆλ, µ, ˆ p) (g

x(x, ˆ p))

(h

x(x, ˆ p))

Λ · gx(x, ˆ p) Γ Θ

hx(x, ˆ p) Θ Θ

v1

v2

v3

= 0nx

0ng

0nh

for v1 ∈ Rnx , v2 = (v21, v22) ∈ Rl+(ng−l), and v3 ∈ Rnh. Exploitation of the special structureof Λ and Γ yields Γ1v21 = 0Rl and since Γ1 is non-singular it follows v21 = 0l. With this, itremains to investigate the reduced system

A B C

B Θ Θ

C Θ Θ

v1

v22

v3

=

0nx

0ng−l

0nh

c 2006 by M. Gerdts

Page 59: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 59/243

3.7. PERTURBED NONLINEAR OPTIMIZATION PROBLEMS 51

with A := Lxx(x, λ, µ, ˆ p), B := (gx,i(x, ˆ p))i=l+1,...,ng , and C := hx(x, ˆ p). Notice, that the second

block equation has been multiplied with Λ−12 . The last two block equations yield Bv1 = 0ng−l

and C v1 = 0nh. Multiplication of the first block equation from the left with v1 yields

0 = v

1 Av

1 + (Bv

1)v

22 + (Cv

1)v

3 = v

1 Av

1.

Since A is positive definite on T C (x) \ 0nx, i.e. it holds dAd > 0 for all d = 0nx withBd = 0ng−l and C d = 0nh

, it follows v1 = 0nx . Taking this property into account, the first blockequation reduces to Bv22 + C v3 = 0nx . By the linear independence of the gradients ∇gi(x, ˆ p),i ∈ A(x, ˆ p) and ∇h j(x, ˆ p), j = 1, . . . , nh we obtain v22 = 0ng−l, v3 = 0nh

. Putting all together,the above linear equation has the unique solution v1 = 0nx , v2 = 0ng , v3 = 0nh

, which impliesthat the matrix F x is non-singular and the implicit function theorem is applicable.By the implicit function theorem there exist neighborhoods V (ˆ p) and U δ(x, λ, µ), and uniquelydefined functions

(x(·), λ(·), µ(·)) : V ε(ˆ p) → U δ(x, λ, µ)

satisfyingF (x( p), λ( p), µ( p), p) = 0nx+ng+nh

(3.7.3)

for all p ∈ V (ˆ p). Furthermore, these functions are continuously differentiable and (3.7.1) arisesby differentiation of the identity (3.7.3) w.r.t. p.It remains to verify, that x( p) actually is a strongly regular local minimum of N LP ( p). Thecontinuity of the functions x( p), λ( p) and g together with λi(ˆ p) = λi > 0, i = l + 1, . . . , ng andgi(x(ˆ p), ˆ p) = gi(x, ˆ p) < 0, i = 1, . . . , l guarantees λi( p) > 0, i = l + 1, . . . , ng and gi(x( p), p) < 0,i = 1, . . . , l for p sufficiently close to ˆ p. From (3.7.3) it follows gi(x( p), p) = 0, i = l + 1, . . . , ng,and h j (x( p), p) = 0, j = 1, . . . , nh. Thus, x( p) ∈ Σ( p) and the KKT conditions are satisfied.Furthermore, the index set A(x( p), p) = A(x, ˆ p) remains unchanged in a neighborhood of ˆ p.Finally, we have to show that L

xx(x( p), λ( p), µ( p), p) remains positive definite on T C (x( p)) for

p sufficiently close to ˆ p. Notice, that the critical cone T C (x( p)) varies with p. By now, we onlyknow that L

xx(ˆ p) := Lxx(x(ˆ p), λ(ˆ p), µ(ˆ p), ˆ p) is positive definite on T C (x(ˆ p)) which is equivalent

to the existence of some α > 0 with dLxx(ˆ p)d ≥ αd2 for all d ∈ T C (x(ˆ p)). Owing to the

strict complementarity in a neighborhood of ˆ p it holds

T C (x( p)) =

d ∈ R

nx

gi(x( p), p)(d) = 0, i ∈ A(x, ˆ p),h j(x( p), p)(d) = 0, j = 1, . . . , nh

around ˆ p. Assume, that for every i ∈ N there exists some p i ∈ Rnp with pi − ˆ p ≤ 1i such that

for all j ∈ N there exists some d ij ∈ T C (x( pi)), dij = 0nx , with

(dij )Lxx( pij)dij < 1

j dij2.

Since the unit ball w.r.t. · is compact in Rnx, there exists a convergent subsequence pijkwith limk→∞ pijk = ˆ p and

limk→∞

dijk

dijk = d, d = 1, d ∈ T C (x(ˆ p))

anddL

xx(ˆ p)d ≤ 0.

This contradicts the positive definiteness of L

xx

(ˆ p).

c 2006 by M. Gerdts

Page 60: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 60/243

52 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Buskens and Maurer [BM01b] exploit Theorem 3.7.3 and equation (3.7.1) to construct an ap-proximation for the optimal solution of a perturbed optimization problem in real-time. Thisreal-time approximation is based on a linearization of the nominal solution for a given nom-inal parameter ˆ p. Under the assumptions of Theorem 3.7.3 the solution x( p) of N LP ( p) iscontinuously differentiable in some neighborhood of ˆ p. Hence, it holds

x( p) = x(ˆ p) + dx

dp(ˆ p)( p − ˆ p) + o( p − ˆ p)

with dx/dp from (3.7.1).

An approximation of the optimal solution x( p) is obtained by the linear approximation

x( p) ≈ x(ˆ p) + dx

dp(ˆ p)( p − ˆ p). (3.7.4)

The time consuming computation of the nominal solution x(ˆ p) and the sensitivity dx/dp is done

offline. The evaluation of the right handside in (3.7.4) is very cheap since only a matrix-vectorproduct and two vector additions are necessary.

The linearization in (3.7.4) is only justified locally in some neighborhood of the nominal pa-rameter ˆ p. Unfortunately, Theorem 3.7.3 does not indicate how large this neighborhood is. Inparticular, if the index set of active inequality constraints changes then Theorem 3.7.3 is notapplicable anymore.

3.8 Numerical Methods

We analyze the Lagrange-Newton-Method and the SQP-Method more closely. The SQP-Methodis discussed in, e.g., [Han77], [Pow78], [GMW81], [Sto85], [Sch81], [Sch83], [Alt02], [GK02]. TheSQP-Method exists in several implementations, e.g. [Sch85], [Kra88], [GMSW98], [GMS02].

Special adaptations of the SQP method to discretized optimal control problems are describedin [GMS94], [Sch96], [Ste95], [BH99]. Again, we restrict the discussion to the finite dimensionalcase. A version of the SQP method working in general Banach spaces can be found in Alt [Alt91].A SQP method for ODE optimal control problems is discussed in Machielsen [Mac88] and forPDE optimal control problems in Troltzsch [Tro05].

3.8.1 Lagrange-Newton-Method

In this section we restrict the discussion to the equality constrained nonlinear optimizationproblem

Problem 3.8.1 Find x ∈ Rnx such that f (x) is minimized subject to the constraints

h j (x) = 0, j = 1, . . . , nh.

The functions f : Rnx → R and h j : Rnx → R, j = 1, . . . , nh, are assumed to be twicecontinuously differentiable. Let x be a local minimum of Problem 3.8.1 and let the gradients∇h j (x), j = 1, . . . , nh be linearly independent. Then the KKT conditions are valid: There existmultipliers µ = (µ1, . . . , µnh

) ∈ Rnh such that

∇xL(x, µ) = ∇f (x) +

nh j=1

µ j∇h j(x) = 0nx ,

h j(x) = 0, j = 1, . . . , nh.

c 2006 by M. Gerdts

Page 61: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 61/243

Page 62: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 62/243

54 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

• The matrix in (3.8.4) is called Kuhn-Tucker-matrix (KT-matrix). The KT-matrix is non-singular, if the following conditions are satisfied:

(i) the gradients ∇h j (x), j = 1, . . . , nh, are linearly independent;(ii) it holds

v

Lxx(x, µ)v > 0

for all 0nx = v ∈ Rnx with h(x) · v = 0nh

.

• The convergence is at least super-linear, if the second derivatives of f and h j , j = 1, . . . , nh,exist: There exists a sequence C k with limk→∞ C k = 0 such that

(x(k+1), µ(k+1)) − (x, µ) ≤ C k(x(k), µ(k)) − (x, µ)

for all sufficiently large k.

3.8.2 Sequential Quadratic Programming (SQP)The linear equation (3.8.2) in item (iii) of the Lagrange-Newton method can be obtained in adifferent way.

Let us again consider the equality constrained problem 3.8.1. We assume, that the problem canbe approximated locally at some point (x(k), µ(k)) by the quadratic optimization problem

mind∈Rnx

1

2dL

xx(x(k), µ(k))d + ∇f (x(k))d

s.t. h(x(k)) + h(x(k))d = 0nh.

The Lagrange function for the quadratic problem is given by

L(d, η) := 12

dLxx(x(k), µ(k))d + f (x(k))d + η h(x(k)) + h(x(k))d .

The evaluation of the first order necessary conditions leads to L

xx(x(k), µ(k)) h(x(k))

h(x(k)) Θ

·

= −

∇f (x(k))

h(x(k))

. (3.8.5)

If we subtract h(x(k))µ(k) on both sides of the first equation in (3.8.5) we get the linear equation L

xx(x(k), µ(k)) h(x(k))

h(x(k)) Θ

·

d

η − µ(k)

= −

∇xL(x(k), µ(k))

h(x(k))

. (3.8.6)

A comparison of (3.8.6) with (3.8.2) reveals that these two linear equation are identical, if weset v := η − µ(k). According to (3.8.3) it follows that the new iterates are given by

x(k+1) = x(k) + d, µ(k+1) = µ(k) + v = η.

Hence, for equality constrained optimization problems, the Lagrange-Newton method is identi-cal with the above depicted successive quadratic programming method, if we use the Lagrangemultiplier η of the QP subproblem as the new approximation for the multiplier µ.

This observation motivates the following extension of the Lagrange-Newton method for Prob-lem 3.6.1 with S = Rnx .

c 2006 by M. Gerdts

Page 63: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 63/243

Page 64: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 64/243

56 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Proof. The proof exploits the sensitivity theorem 3.7.3 for parametric optimization problems toshow that the index set of active constraints remains unchanged within a certain neighborhoodof the solution. Then, it is possible to show that the SQP method locally coincides with theLagrange-Newton method.

(a) We consider QP (x, λ, µ) as the unperturbed quadratic problem with nominal parameterˆ p = (x, λ, µ). Furthermore, we notice, that the KKT conditions for QP (ˆ p) and Prob-lem 3.6.1 with S = Rnx coincide for d = 0nx . Hence, (0nx , λ, µ) is a KKT point of QP (ˆ p).In addition, assumptions (iii)-(v) guarantee that d is a strongly regular local minimum of QP (ˆ p).

Hence, we may apply Theorem 3.7.3: There exists a neighborhood V ε(ˆ p) such that QP ( p)has a unique strongly regular local minimum (d( p), λ( p), µ( p)) for each p ∈ V ε(ˆ p). Herein,(d( p), λ( p), µ( p)) is continuously differentiable w.r.t. p.

Furthermore, the index set of active constraints remains unchanged in that neighborhood:A(x) = AQP (d, ˆ p) = AQP (d( p), p). AQP (d, p) denotes the index set of active inequality

constraints of QP ( p) at d.

(b) Due to the continuity of the constraints, we may neglect the inactive constraints at x andobtain the (locally) equivalent optimization problem

min f (x) s.t. gi(x) = 0, i ∈ A(x), h j(x) = 0, j = 1, . . . , nh.

We can apply the Lagrange-Newton method and under the assumptions (i)-(v) Theo-rem 3.8.3 yields the local quadratic convergence

(x(k), λ(k)A(x)

, µ(k)) → (x, λA(x), µ).

Notice, that the multipliers are unique according to (iii). We may add λ(k)i = 0 for i ∈ A(x)

to obtain

p(k) := (x(k), λ(k), µ(k)) → ˆ p = (x, λ, µ).

(c) Let δ denote the radius of convergence of the Lagrange-Newton method. Let r := minε, δ .For p(0) ∈ U r(ˆ p) all subsequent iterates p(k) of the Lagrange-Newton method remain inthat neighborhood.

Furthermore, (d(k), λ(k+1), µ(k+1)) with d(k) = x(k+1) −x(k) fulfills the necessary conditionsof QP ( p(k)), cf. (3.8.5) and (3.8.6). According to (a), the solution (d( p(k)), λ( p(k)), µ( p(k)))of QP ( p(k)) is unique. Hence, the SQP iteration coincides with the Lagrange-Newtoniteration.

Remark 3.8.8 (Approximation of Hessian)The use of the exact Hessian L

xx of the Lagrange function in the QP problem has two drawbacks from numerical point of view:

• In most practical applications the Hessian is not known explicitly. The numerical approx-imation of the Hessian by finite differences is very expensive.

• The Hessian may be indefinite. This makes the numerical solution of the QP problem more difficult. It is desirable to have a positive definite matrix in the QP problem.

c 2006 by M. Gerdts

Page 65: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 65/243

3.8. NUMERICAL METHODS 57

In practice, the Hessian of the Lagrange function in iteration k is replaced by a suitable matrix Bk. Powell [Pow78] suggested to use the modified BFGS-update formula

Bk+1 = Bk + q (k)(q (k))

(q (k))s(k) − Bks(k)(s(k))Bk

(s(k))Bks(k)

, (3.8.7)

where

s(k) = x(k+1) − x(k),

q (k) = θkη(k) + (1 − θk)Bks(k),

η(k) = ∇xL(x(k+1), λ(k), µ(k)) − ∇xL(x(k), λ(k), µ(k)),

θk =

1, if (s(k))η(k) ≥ 0.2(s(k))Bks(k),

0.8(s(k))Bks(k)

(s(k))Bks(k)−(s(k))η(k), otherwise .

This update formula guarantees that Bk+1 remains symmetric and positive definite if Bk was symmetric and positive definite. For θk = 1 we get the well known BFGS update formula, which is used in variable metric methods (or quasi Newton methods) for unconstrained optimization.

If the exact Hessian is replaced by the modified BFGS update formula, the convergence of the resulting SQP method is only super-linear.

3.8.3 Globalization of the Local SQP Method

The convergence result shows that the SQP method converges for all starting values which arewithin some neighborhood of a local minimum of Problem 3.6.1 with S = Rnx . Unfortunately,in practice this neighborhood is not known and it cannot be guaranteed, that the starting values

are within this neighborhood. Fortunately, the SQP method can be globalized in the sense thatit converges for arbitrary starting values (under suitable conditions). The idea is to determinethe new iterate x(k+1) according to the formula

x(k+1) = x(k) + tkd(k)

with a step length tk > 0. The step length tk is obtained by performing a so-called line search inthe direction d(k) for a suitable penalty function or merit function . The penalty function allowsto decide whether the new iterate x(k+1) is in some sense ‘better’ than the old iterate x(k).The new iterate will be better than the old one, if either a sufficient decrease in the objective

function f or an improvement of the total constraint violations is achieved while the respectiveother value is not substantially declined.

Often, one of the following merit functions is used for globalization:

• The non-differentiable 1-penalty function

1(x; α) := f (x) + α

ngi=1

max0, gi(x) + α

nh j=1

|h j(x)|

was used by, e.g., Powell [Pow78].

c 2006 by M. Gerdts

Page 66: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 66/243

58 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

• A commonly used differentiable penalty function is the augmented Lagrange function

La(x,λ,µ; α) = f (x) + µh(x) + α

2h(x)2

+

1

ngi=1

(max0, λi + αgi(x))2

− λ2

i= f (x) +

nh j=1

µ jh j (x) +

α

2h j (x)2

(3.8.8)

+

ngi=1

λigi(x) + α

2 gi(x)2, if λi + αgi(x) ≥ 0,

− λ2i

2α , otherwise.

A SQP method employing the augmented Lagrange function is discussed in [Sch81],[Sch83].

Both functions are exact under suitable assumptions, i.e. there exists a finite parameter α > 0,such that every local minimum x of Problem 3.6.1 with S = Rnx is also a local minimum of thepenalty function for all α ≥ α.Without specifying all details, a globalized version of the SQP method employing the 1 penaltyfunction and an Armijo rule for step length determination reads as follows.

Algorithm 3.8.9 (Globalized SQP Method)

(i) Choose (x(0), λ(0), µ(0)) ∈ Rnx ×Rng ×Rnh, B0 ∈ Rnx×nx symmetric and positive definite,α > 0, β ∈ (0, 1), σ ∈ (0, 1), and set k = 0.

(ii) If (x(k), λ(k), µ(k)) is a KKT point of Problem 3.6.1 with S = Rnx, STOP.

(iii) Compute a KKT point (d(k), λ(k+1), µ(k+1)) ∈ Rnx×Rng×Rnh of the quadratic programming problem QP (x(k), λ(k), µ(k)) with L

xx replaced by Bk.

(iv) Determine a step size tk = maxβ j | j = 0, 1, 2, . . . such that

1(x(k) + tkd(k); α) ≤ 1(x(k); α) + σtk1(x(k); d(k); α).

(v) Compute Bk+1 according to (3.8.7) and set x(k+1) := x(k) + tkd(k).

(vi) Set k := k + 1 and go to (ii).

Remark 3.8.10

• In practical applications a suitable value for the penalty parameter α is not known a priori.Strategies for adapting α iteratively and individually for each constraint can be found in [Sch83] and [Pow78].

• So far, we always assumed that the QP problem has a solution. This assumption is not always justified, even if the original problem is feasible. To overcome this problem, Pow-ell [Pow78] suggested to relax the constraints of the QP problem in such a way, that the relaxed QP problem possesses admissible points. The infeasible QP problem is then re-placed by the feasible relaxed problem. A convergence analysis for a SQP method using the augmented Lagrangian can be found in Schittkowski [Sch81, Sch83].

c 2006 by M. Gerdts

Page 67: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 67/243

3.8. NUMERICAL METHODS 59

3.8.4 Nonsmooth Newton’s Method for Quadratic Programs

One approach to solve the quadratic program is to employ primal or dual active set methodsas in Gill et al. [GM78, GMSW91] or Goldfarb and Idnani [GI83]. The alternative approach of Lemke [Lem62] uses pivot strategies to solve the linear complementarity problem resulting fromthe KKT conditions. We will discuss a third strategy.For simplicity we consider quadratic programs (QP) of type

minx∈Rn

1

2xQx + cx s.t. g(x) := Ax − u ≤ 0m,

where Q ∈ Rn×n is symmetric and positive definite, A ∈ Rm×n, c ∈ Rn, and u ∈ Rm. Additionalequality constraints can be easily added and will not effect the subsequent analysis.Let x∗ be a local minimum of QP. Then, the first order necessary Karush-Kuhn-Tucker (KKT)conditions read as follows: There exists a multiplier λ ∈ Rm such that

Qx + c + Aλ = 0n, λ ≥ 0, λg(x) = 0, g(x) ≤ 0m. (3.8.9)

The idea is to restate the KKT conditions (3.8.9) as the equivalent nonlinear equation

F (z) = 0n+m (3.8.10)

with variables z = (x, λ) and

F (z) :=

Qx + c + Aλϕ(−g1(x), λ1)

...ϕ(−gm(x), λm)

= 0n+m. (3.8.11)

Herein, the function ϕ : R2

→R is a so-called NCP function with the property

ϕ(a, b) = 0 ⇔ a ≥ 0, b ≥ 0, a · b = 0.

Obviously, the KKT conditions (3.8.9) are satisfied if and only if (3.8.11) holds. It is easy tocheck that the convex Fischer-Burmeister function

ϕ(a, b) :=

a2 + b2 − a − b.

is a particular NCP function. Unfortunately, the Fischer-Burmeister function ϕ is not differ-entiable at the origin (a, b) = (0, 0). Hence, the function F in (3.8.10) is not differentiable atpoints with (gi(x), λi) = (0, 0) and the common Newton’s method is not applicable to the non-linear equation (3.8.11). In the sequel we will discuss a nonsmooth version of Newton’s method

adapted to the particular equation in (3.8.11). Nonsmooth Newton’s method for general equa-tions are discussed in Qi [Qi93], Qi and Sun [QS93], Jiang [Jia99], Xu and Glover [XG97], Xu andChang [XC97], Han et al. [HPR92], Ralph [Ral94], and Dingguo and Weiwen [DW02]. Similarideas are also used to solve nonlinear complementarity problems, cf. Fischer [Fis97], Facchineiand Kanzow [FK97], Yamashita and Fukushima [YF97] and Billups and Ferris [BF97].The components of F are convex functions (gi is linear, ϕ is convex) and thus F is locallyLipschitz continuous. According to Rademacher’s theorem, F is differentiable almost everywhereand we may define the B(ouligand)-differential

∂ BF (z) :=

V

V = lim

zi∈DF z

i

→z

F (zi)

,

c 2006 by M. Gerdts

Page 68: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 68/243

60 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

where DF denotes the set of points where F is differentiable. Notice, that Clarke’s [Cla83]generalized Jacobian ∂ F is the convex hull of the B-differential:

∂F (z) := conv(∂ BF (z)).

The generalized Jacobian ∂F (z) is a non-empty, convex, and compact set and as a function of z it is upper semicontinuous at z , cf. Clarke [Cla83], Prop. 2.6.2, p. 70.

Definition 3.8.11 (Qi [Qi93])Let F : Rn → Rm.

• ∂F and ∂ BF , respectively, are called non-singular at z, if every V ∈ ∂F (z) resp. every V ∈ ∂ BF (z) is non-singular.

• F is called B(ouligand)-differentiable at x if it is directionally differentiable, i.e. the limit

F (x; h) = limt↓0

F (x + th)−

F (x)

t

exists for every direction h, and it holds

limh→0

F (x + h) − F (x) − F (x; h)

h = 0

for every direction h.

• Let F be B-differentiable in some neighborhood of x. F (·; ·) is semicontinuous at x, if for every ε > 0 there exists a neighborhood N of x such that

F (x + h; h) − F (x; h) ≤ εh ∀x + h ∈ N.

F (·; ·) is semicontinuous of degree 2 at x, if there exist a constant L and a neighborhood N of x such that

F (x + h; h) − F (x; h) ≤ Lh2 ∀x + h ∈ N.

• F is called semismooth at x, if

limV ∈∂F (x+th)

h→h,t↓0

V h

exists for every h ∈Rn

.We summarize results from Qi [Qi93] and Qi and Sun [QS93]. In finite dimensional spacesShapiro showed that a locally Lipschitz continuous function is B-differentiable if and only if itis directionally differentiable.

Lemma 3.8.12 (Qi [Qi93], Qi and Sun [QS93])

Let F : Rn → Rm be directionally differentiable in some neighborhood of x.

(a) The following statements are equivalent:

– F is semismooth at x.– F (

·;·) is semicontinuous at x.

c 2006 by M. Gerdts

Page 69: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 69/243

3.8. NUMERICAL METHODS 61

– For every V ∈ ∂F (x + h), h → 0 it holds

V h − F (x; h) = o(h).

– limx+h∈DF h→0

F (x + h; h)

−F (x; h)

h = 0.

(b) The following statements are equivalent:

– F (·; ·) is semicontinuous of degree 2 at x.– For every V ∈ ∂F (x + h), h → 0 it holds

V h − F (x; h) = O(h2). (3.8.12)

In either case F is B-differentiable of degree 2 at x.

In particular, convex functions and smooth functions are semismooth. Scalar products, sums,and compositions of semismooth functions are semismooth. The min-Operator is semismooth.

Local Convergence

First, we investigate a local nonsmooth Newton’s method and its convergence.

Algorithm 3.8.13 (Local Nonsmooth Newton’s Method)

(0) Choose z0 and set k = 0.

(1) If some stopping criterion is satisfied, stop.

(2) Choose V k ∈ ∂ BF (zk) and compute the search direction dk as the solution of the linear equation

V kd = −F (zk).

(3) Set zk+1 = zk + dk, k = k + 1, and goto (1).

The following convergence theorem can be found in Qi [Qi93].

Theorem 3.8.14 (Qi [Qi93], Th. 3.1)Let z∗ satisfy (3.8.10). Let F be locally Lipschitz continuous and semismooth at z∗ and let

∂ BF (z∗) be nonsingular. Then, there exists some r > 0 such that for any z0 ∈ U r(z∗) Algo-rithm 3.8.13 is well-defined and the sequence zk converges superlinearly to z∗.

If F (zk) = 0 for all k, then

limk→∞

F (zk+1)F (zk) = 0.

If in addition F is directionally differentiable in a neighborhood of z∗ and F (·; ·) is semicontin-uous of degree 2 at z∗, then the convergence is quadratic.

Qi and Sun [QS93], Th. 3.2 proved a similar result, if V k is chosen from ∂F (zk) in step (2) of Algorithm 3.8.13.

c 2006 by M. Gerdts

Page 70: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 70/243

62 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Global Convergence

One reason that makes the Fischer-Burmeister function appealing is the fact that its square

ψ(a, b) := ϕ(a, b)2

= a2

+ b2

− a − b2

is continuously differentiable. This allows to globalize the local Newton’s method by introducingthe merit function

Ψ(z) := 1

2F (z)2 =

1

2F (z)F (z) =

1

2

n+mi=1

F i(z)2

= 1

2

Qx + c + Aλ2 +

mi=1

ψ(−gi(x), λi)

.

Ψ is continuously differentiable and it holds

∇Ψ(z) = V F (z),

where V is an arbitrary element of ∂F (z).

Algorithm 3.8.15 (Globalized Nonsmooth Newton’s Method)

(0) Choose z0, β ∈ (0, 1), σ ∈ (0, 1/2) and set k = 0.

(1) If some stopping criterion is satisfied, stop.

(2) Compute the search direction dk as the solution of the linear equation

V kd = −F (zk),

where V k ∈ ∂F (zk).

(3) Find the smallest ik ∈ N0 with

Ψ(zk + β ikdk) ≤ Ψ(zk) + σβ ik∇Ψ(zk)dk

and set αk = β ik .

(4) Set zk+1 = zk + αkdk, k = k + 1, and goto (1).

The following result is Theorem 4.1 in Jiang [JQ97].

Theorem 3.8.16 Let the assumptions of Theorem 3.8.14 be satisfied. Let the linear equation in step (2) be solvable and let z∗ be an accumulation point of the sequence zk generated by Algorithm 3.8.15.

Then:

(i) z∗ is a zero of F if dk is bounded.

(ii) z∗ is a zero of F and zk converges to z∗ superlinearly if σ ∈ (0, 1/2) and if ∂F (z∗) is nonsingular.

c 2006 by M. Gerdts

Page 71: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 71/243

3.8. NUMERICAL METHODS 63

There are two assumptions that ought to be discussed. The first one is the nonsingularity of ∂F (z). The second one is the existence of an accumulation point. An accumulation point exists,if the sequence zk is bounded. This would be the case, if the level set

z |

Ψ(z) ≤

Ψ(z0)

is bounded. Unfortunately, the question of whether the level set is bounded or not remains open.The problem is, that any point (x, λ) satisfying the complementarity conditions −gi(x) ≥ 0,λi ≥ 0, λigi(x) = 0 yields ϕ(−gi(x), λi) = 0. As a consequence, it may happen that a sequence(xk, λk) with (xk, λk) → ∞ and Ψ(xk, λk) < ∞ exists.

Lemma 3.8.17 The Fischer-Burmeister function ϕ is both, semismooth and semicontinuous of degree 2.

Proof. For (a, b) = (0, 0), ϕ is smooth in some neighborhood and thus

ϕ

(a, b; h1, h2) = ϕ

(a, b)h.

For (a, b) = (0, 0), ϕ is directionally differentiable with

ϕ(0, 0; h1, h2) = limt↓0

ϕ(th1, th2) − ϕ(0, 0)

t = lim

t↓0

t2h2

1 + t2h22 − th1 − th2

t = ϕ(h1, h2).

Semismoothness: We show the equivalent characterizations

V h − ϕ(a, b; h1, h2) = o(h) ∀V ∈ ∂ϕ(a + h1, b + h2) as h → 0

for semismoothness and

V h − ϕ(a, b; h1, h2) = O(h2) ∀V ∈ ∂ϕ(a + h1, b + h2) as h → 0.

for semicontinuity of degree 2.

Let (a, b) = (0, 0). Then since ϕ is twice continuously differentiable in a neighborhood of (a, b):

|V h − ϕ(a, b; h1, h2)| = |ϕ(a + h1, b + h2)h − ϕ(a, b)h|≤ ϕ(a + h1, b + h2) − ϕ(a, b)

h≤ Lh2.

Let (a, b) = (0, 0) and h = 0. Then

V h − ϕ(0, 0; h1, h2) = ϕ(h1, h2)h − ϕ(h1, h2) = h2

1

h − h1 + h2

2

h − h2 − h + h1 + h2 = 0.

This proves the assertion.

The chain law for directionally differentiable functions reads as follows, cf. Geiger and Kan-zow [GK02], Theorem 5.33, p. 251: Given three functions h : Rn → Rm, g : Rm → R p, andf = g h, where h and g are directionally differentiable at x resp. h(x) and g is locally Lipschitzcontinuous at h(x). Then f is directionally differentiable at x with

f (x; d) = g (h(x); h(x; d)).

c 2006 by M. Gerdts

Page 72: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 72/243

64 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Hence, application of this result to our situation yields that F is also semismooth and F issemicontinuous of degree 2.

Nonsingularity of Subdifferential

We first start with the investigation of the subdifferential of ϕ. If (a, b) = (0, 0), then ϕ isdifferentiable and

ϕ(a, b) =

a√ a2 + b2

− 1, b√ a2 + b2

− 1

.

In order to obtain the B-subdifferential we have to investigate all limits of ϕ(a, b) for (a, b) →(0, 0). It holds

ϕ(a, b) + (1, 1) ≤ 1 ∀(a, b) = (0, 0)

and thus

ϕ(a, b) ∈ M := (s, t) | (s + 1)2 + (t + 1)2 ≤ 1 ∀(a, b) = (0, 0).

Consequently,∂ Bϕ(0, 0) ⊆ ∂ϕ(0, 0) ⊆ M.

On the other hand, using the sequences (ai, bi) = 1i (cos α, sin α) with arbitrary α ∈ [0, 2π) it

follows

limi→∞

ϕ(ai, bi) = (cos α − 1, sin α − 1) .

Hence,

∂ϕ(0, 0) = M.

Any element from ∂ Bϕ(0, 0) resp. ∂ϕ(0, 0) can be represented as (s, t) with (s+1)2 +(t+1)2 ≤ 1.On the other hand, we did not show that every such (s, t) is actually an element of ∂ Bϕ(0, 0).

Summarizing, we obtained

∂ϕ(a, b) =

a√ a2 + b2

− 1, b√ a2 + b2

− 1

, if (a, b) = (0, 0),

(s, t) (s + 1)2 + (t + 1)2 ≤ 1

, if (a, b) = (0, 0).

For the forthcoming computations we used the particular subgradient with s = 0, t = 1, whichis an element of ∂ Bϕ(0, 0).

The first n components of F in (3.8.11) are continuously differentiable. The last m componentsof F in (3.8.11) are composite functions of the locally Lipschitz continuous function ϕ and the

affine function (x, λ) → −gi(x)

λi . Hence we need a chain rule for differentiation. Application

of Theorem 2.6.6 in Clarke [Cla83] yields

∂F n+i(x, λ) ⊆ co

∂ϕ(−gi(x), λi) ·

−ai 00 ei

for i = 1, . . . , m, where ai denotes the i-th row of A and ei the i-th unity vector of dimension Rm.The convex hull can be omitted since from convex analysis it is well-known that the product of aconvex set by a matrix is again a convex set, i.e. if C is convex and A is a matrix of appropriatedimension, then C · A is convex. Hence, for i = 1, . . . , m it holds

∂F n+i(x, λ)

⊆∂ϕ(

−gi(x), λi)

· −ai 0

0 ei .

c 2006 by M. Gerdts

Page 73: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 73/243

3.8. NUMERICAL METHODS 65

Using Proposition 2.6.2 of Clarke [Cla83] we finally find

∂F (x, λ) ⊆

Q A

−SA T

,

where S := diag(s1, . . . , sm), T := diag(t1, . . . , tm), and (si, ti) ∈ ∂ϕ(−gi(x), λi). In particular,

(si + 1)2 + (ti + 1)2 ≤ 1,

cf., eg., Jiang [Jia99]. Now, we will show that any such matrix is nonsingular, if Q is positivedefinite and if the constraints do not contain redundant information. Define

α(x) := i ∈ 1, . . . , m | gi(x) = 0.

Theorem 3.8.18 Let Q be positive definite and let the constraints Ax ≤ u be regular in the sense that for any x ∈ Rn the rows ai | i ∈ α(x) are linearly independent. Then, ∂F (x, λ) is nonsingular for any (x, λ).

Proof. Any element from ∂ F (x, λ) can be represented by

V :=

Q A

−SA T

,

where S := diag(s1, . . . , sm), T := diag(t1, . . . , tm), and

(si + 1)2 + (ti + 1)2 ≤ 1,

Define

I := i ∈ 1, . . . , m | si = 0,

J := i ∈ 1, . . . , m | ti = 0,K := i ∈ 1, . . . , m | i ∈ I ∪ J .

Notice, that I ∩ J = ∅ because of (si + 1)2 + (ti + 1)2 ≤ 1. In particular it holds ti = 0 for i ∈ I ,si = 0 for i ∈ J , and s i, ti ∈ [−2, 0) for i ∈ K .

In order to show the nonsingularity of V we show that the linear equation V

uv

= 0n+m

only admits the trivial solution. So, let

Qu + Av = 0n,

−SAu + T v = 0m

resp.

Qu +i∈I

ai vi + j∈J

a j v j +k∈K

ak vk = 0n+m,

−siaiu + tivi = 0, i ∈ I ,

−s ja ju + t jv j = 0, j ∈ J,

−skaku + tkvk = 0, k ∈ K.

Since si = 0, ti = 0 for i ∈ I the second equality immediately yields vi = 0 for i ∈ I . Similarly,the third equation yields a ju = 0 for j ∈ J because s j = 0, t j = 0 for j ∈ J . In particular,v ja j u = 0, j

∈ J .

c 2006 by M. Gerdts

Page 74: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 74/243

66 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Multiplication of the last equation by v k yields

tk

skv2

k = vkaku,

where tk/sk > 0. Recall that sk, tk ∈ [−2, 0).

Multiplication of the first equation and exploitation of the above relations yields

0 = uQu +i∈I

uai vi + j∈J

ua j v j +k∈K

uak vk = uQu +k∈K

tk

skv2

k.

Due to the positive definiteness of Q this equality only holds for u = 0 and v k = 0, k ∈ K .By now, we obtained u = 0, vi = 0, i ∈ I , and vk = 0, k ∈ K . With this, again the first equationyields

j∈J

a j v j = 0.

Recall that j ∈ J means t j = 0 and (s j, t j) ∈ ∂ϕ(−g j(x), λ j ) and (s j + 1)2 ≤ 1. The value t j = 0is only possible, if g j (x) = a j x

−u j = 0 holds, i.e. j

∈ α(x). According to our assumption the

set a j | j ∈ α(x) is linearly independent. Thus, v j = 0, j ∈ J . This completes the proof.

Example 3.8.19 We will test the performance of the nonsmooth Newton’s method on two ran-domly generated test sets with strictly convex objective function. In addition, we will comparethe results in view of iterations needed with those obtained by the active set method quadprog

from Matlab. It has to be mentioned, that the respective methods approximately need thesame amount of time per iteration, provided that the structure of the occurring linear equationsis exploited.

• The first test set consists of 941 small scale problems. The problem sizes range from 2 to20 variables and from 1 to 20 constraints.

• The second test set consists of 261 medium to large scale problems. The problem sizesrange from 2 to 452 variables and from 1 to 801 constraints.

Table 3.1 summarizes the results of the active set method and the nonsmooth Newton’s methodfor the first test set with few constraints. The active set method always needed less iterations.

Table 3.1: Primal active set method vs. nonsmooth Newton’s method: Iterations for problemswith few constraints.

method average min max

active set 5 1 25newton 8 2 37

Table 3.2 summarizes the results of the active set method and the nonsmooth Newton’s methodfor the second test set with many constraints. The nonsmooth Newton’s method performssignificantly better for many constraints than the active set method. This suggests that thenonsmooth Newton’s method might be superior for problems with many constraints.

Table 3.2: Primal active set method vs. nonsmooth Newton’s method: Iterations for problemswith many constraints.

method average min max

active set 197 1 987newton 12 2 56

c 2006 by M. Gerdts

Page 75: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 75/243

3.8. NUMERICAL METHODS 67

Example 3.8.20 (Discretized Elliptic PDE Optimal Control Problem)We consider the subsequent elliptic PDE Optimal Control Problem with an objective functionof tracking type:

miny,u∈L2(Ω)

1

2y − yd2

L2(Ω) + α

2u2

L2(Ω)

subject to the elliptic PDE

−∆y = β · u in Ω := (0, 1) × (0, 1),y = 0 on Γ = ∂ Ω

and the box constraintsb(x) ≤ u(x) ≤ bu(x).

The problem is discretized using the 5-point star approximation

−∆y(xij) ≈ 4yij − yi−1,j − yi,j−1 − yi+1,j − yi,j+1

h2

on the equidistant grid xij = (ih,jh), i, j = 0, . . . , N , h = 1/N with approximations yij ≈ y(xij ),uij ≈ u(xij). The integral is approximated by

Ωy(x)dx ≈ h2

ij

yij.

The discretized problem reads as

min 1

2

N −1i,j=1

(yij − yd(xij))2 + αu2

ij

subject toAhy = Bhu, b(xij ) ≤ uij ≤ bu(xij),

where Ah is a banded matrix of bandwidth N with approximately 5/(N − 1)% of nonzeroelements. Bh is a diagonal matrix with entries β (xij ).The discretized problem is equivalent to the quadratic program

minz

1

2zQz + cz s.t. Aeqz = 0, Ainz ≤ b

where

Q = I Θ

Θ αI , Aeq = Ah

−Bh , Ain = Θ I

Θ −I and z = (yij, uij ) ∈ R2(N −1)2 , c = (−yd(xij), 0, . . . , 0), b = (bu(xij ), −bl(xij ))) . The matri-ces Q, Aeq, and Ain are large but sparse. Hence, instead of treating the Newton equation

V kd = −F (xk)

by dense lu-decomposition a conjugate gradient method is applied. Actually, we solved

V k V kd = −V k F (xk)

using the cg-method. Since V k is nonsingular, both equations are equivalent but V k V k is positivedefinite.

c 2006 by M. Gerdts

Page 76: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 76/243

68 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

The following numerical results are obtained for the data β ≡ 1, b ≡ −200, bu ≡ 200, α = 10−6,N = 60, and

yd(x1, x2) = 1, if (x1 − 0.7)2 + (x2 − 0.5)2 ≤ 0.22,

−1, if (x1

−0.3)2 + (x2

−0.5)2

≤ 0.22,

0, otherwise

The resulting QP problem has 6962 variables, 3481 equality constraints, and 6962 inequalityconstraints. The nonsmooth Newton’s method has 17405 equations, 209045 nonzero elementsin the Jacobian, i.e. ≈ 0.07 % nonzero elements.

-1.5-1-0.5 0 0.5 1 1.5

state

0 0.2

0.4 0.6

0.8

x_1

0.2

0.4 0.6

0.8 1

x_2

-1.5-1

-0.50

0.51

1.5

-1.5-1-0.5 0 0.5

1 1.5 2

multiplier pde

0 0.2

0.4 0.6

0.8x_1

0.2 0.4

0.6 0.8

1

x_2

-1.5-1

-0.500.51

1.52

c 2006 by M. Gerdts

Page 77: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 77/243

3.8. NUMERICAL METHODS 69

-250-200-150-100-50050100150200250

control

0 0.2

0.4

0.6 0.8x_1 0.2 0.4

0.6 0.8

1

x_2

-250-200-150-100-50050100150200250

-5e-05 0 5e-05 0.0001

0.00015 0.0002 0.00025

multiplier upper bound

0 0.2

0.4 0.6

0.8x_1

0.2 0.4

0.6 0.8

1

x_2

-5e-050

5e-050.00010.00015

0.00020.00025

c 2006 by M. Gerdts

Page 78: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 78/243

70 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

-5e-05 0 5e-05 0.0001 0.00015 0.0002 0.00025

multiplier lower bound

0 0.2

0.4

0.6 0.8x_1 0.2 0.4

0.6 0.8

1

x_2

-5e-050

5e-050.0001

0.000150.0002

0.00025

3.9 Duality

We consider Problem 3.6.1 and refer to it as the primal optimization problem . In contrast tothe previous discussions, we do not assume that the functions f, g, and h are continuous or evendifferentiable. Similarly, the set S ⊆ Rnx may be an arbitrary non-empty set.In order to derive the dual problem we consider the

Problem 3.9.1 (Perturbed Optimization Problem)Find x ∈ Rnx such that f (x) is minimized subject to the constraints

gi(x) ≤ yi, i = 1, . . . , ng,

h j (x) = z j, j = 1, . . . , nh,

x ∈ S.

Herein, y = (y1, . . . , yng ) ∈ Rng and z = (z1, . . . , znh) ∈ Rnh are arbitrary vectors. Notice,

that the original (unperturbed) problem arises for y = 0ng and z = 0nh.

Definition 3.9.2 (Minimum Value Function)

The function Φ : Rng+nh → R := R ∪ ∞, −∞

with

Φ(y, z) := inf f (x) | gi(x) ≤ yi, i = 1, . . . , ng, h j (x) = z j, j = 1, . . . , nh, x ∈ S

is called minimum value function for the perturbed problem.

The dual problem is motivated by the task of approximating the graph of the minimum valuefunction Φ from below by a hyperplane

λy + µz + r = γ (3.9.1)

c 2006 by M. Gerdts

Page 79: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 79/243

3.9. DUALITY 71

r

Φ(y, z)

0

(y ,z ,r) ∈Rng+nh+1 | λy + µz + r = γ

Rng+nh

(−λ, −µ)

(−λ, −µ, −1)

primal and dual solution

Figure 3.4: Graphical interpretation of the dual problem: Approximation of the graph of theminimum value function from below by a hyperplane with normal vector (λ,µ, 1) respectively(−λ, −µ, −1).

with normal vector (λ,µ, 1) ∈ Rng+nh+1 and variables (y ,z ,r) ∈ Rng+nh+1, compare Fig-ure 3.4. The intersection of this hyperplane with the r-axis at y = 0ng , z = 0nh

is (0ng , 0nh, γ ).

The dual problem is given by

Maximize γ w.r.t. λ ∈ Rng , µ ∈Rnh

s.t. r ≤ Φ(y, z) ∀y ∈Rng , z ∈ Rnh.

where r from (3.9.1) is given by r = γ −λy−µz. Hence, the dual problem can be reformulatedas

Maximize γ w.r.t. λ ∈ Rng , µ ∈ Rnh

s.t. γ − λy − µz ≤ Φ(y, z) ∀y ∈ Rng , z ∈ Rnh.

By the definition of Φ we get the equivalent problem

Maximize γ w.r.t. λ ∈ Rng , µ ∈Rnh

s.t. γ ≤ f (x) + λy + µz ∀y ∈ Rng , z ∈ Rnh, x ∈ S, g(x) ≤ y, h(x) = z.

Since h(x) = z this problem is equivalent to

Maximize γ w.r.t. λ ∈ Rng , µ ∈Rnh

s.t. γ ≤ f (x) + λy + µh(x) ∀y ∈ Rng , x ∈ S, g(x) ≤ y.

We distinguish two cases. First, we investigate the constraint

γ ≤

f (x) + λy + µh(x) ∀

y ∈R

ng , x∈

S, g(x) ≤

y (3.9.2)

c 2006 by M. Gerdts

Page 80: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 80/243

Page 81: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 81/243

3.9. DUALITY 73

• Consider the primal quadratic optimization problem

min 1

2xQx + cx s.t. Ax ≤ b

with symmetric and positive definite matrix Q. The dual objective function is given by

ψ(λ) = inf x∈Rnx

1

2xQx + cx + λ(Ax − b)

.

The infimum can be calculated: A necessary and sufficient condition is

Qx + Aλ + c = 0, resp. x = −Q−1(c + Aλ).

With the abbreviations

W = −AQ−1A, v = −b − AQ−1c

we obtain the dual problem

maxλ≥0

1

2λW λ + λv − 1

2cQ−1c

.

The advantage of the dual problem is the simpler structure of the constraints (only signconditions).

The following theorem can be exploited for Branch & Bound methods to construct lower boundsfor the primal problem.

Theorem 3.9.5 (Weak duality theorem)Let x be admissible for the primal problem and (λ, µ) be admissible for the dual problem. Then the minimal value w(P ) of the primal problem and the maximal value w(D) of the dual problem satisfy

ψ(λ, µ) ≤ w(D) ≤ w(P ) ≤ f (x).

Proof. It holds

ψ(λ, µ) = inf z∈S

L(z,λ,µ) ≤ f (x) + λ ≥0ng

g(x) ≤0ng

+µ h(x) =0nh

≤ f (x).

The left side is independent of x, while the right side is independent of λ, µ. Taking the supremumw.r.t. λ ≥ 0ng , µ on the left and the infimum w.r.t. x ∈ S, g(x) ≤ 0 ng , h(x) = 0nh

on the right

proves the statement.

If the primal objective function value for some primal feasible solution equals the dual objectivefunction value for some dual feasible solution, then both problems are solved optimally already,because we have

Theorem 3.9.6 (Sufficient optimality criterion)Let ψ(λ, µ) = f (x), where x is admissible for the primal problem and (λ, µ) is admissible for the dual problem. Then x is optimal for the primal problem and (λ, µ) is optimal for the dual problem. In addition, the complementary slackness condition holds:

λi = 0, if gi(x) < 0 (i = 1, . . . , ng).

c 2006 by M. Gerdts

Page 82: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 82/243

74 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Proof. The first statement is a direct consequence of the weak duality theorem. Comple-mentary slackness condition: Assume that gi(x) < 0 and λi > 0 for at least one 1 ≤ i ≤ ng.Then

ψ(λ, µ) = inf x∈S

L(x, λ, µ) ≤ f (x) + λg(x)

<0

+ µh(x)

=0

< f (x)

in contradiction to the assumption ψ(λ, µ) = f (x).

The following example shows, that the case w(D) < w(P ) in fact occurs. In this case a duality gap occurs, compare Figure 3.5.

r

Φ(y, z)

0

(y ,z ,r) ∈Rng+nh+1 | λy + µz + r = γ

Rng+nh

(−λ, −µ)

(−λ, −µ, −1)

duality gap

Figure 3.5: Graphical interpretation of the dual problem: Approximation of the graph of theminimum value function from below by a hyperplane with normal vector (λ,µ, 1) respectively(−λ, −µ, −1) in the presence of a duality gap.

Example 3.9.7 Duality gap (cf. [BS79], p. 181)

Minimize f (x1, x2) = −2x1 + x2

s.t. (x1, x2) ∈ S = (0, 0), (0, 4), (4, 4), (4, 0), (1, 2), (2, 1),0 = h(x1, x2) = x1 + x2 − 3.

The solution of the primal problem is (2, 1) with f (2, 1) = −3. The dual objective function is given by

ψ(µ) = min−2x1 + 3 + µ(x1 + x2 − 3) | (x1, x2) ∈ S =

−4 + 5µ, if µ ≤ −1,−8 + µ, if − 1 ≤ µ ≤ 2,−3µ, if µ ≥ 2.

The optimal solution of the dual problem maxµ∈R ψ(µ) is given by µ = 2 with value −6. Since

−6 <

−3 there exists a duality gap.

c 2006 by M. Gerdts

Page 83: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 83/243

3.10. MIXED-INTEGER NONLINEAR PROGRAMS AND BRANCH&BOUND 75

There remains the question under which conditions a duality gap does not occur. The followingtheorem states a sufficient condition.

Theorem 3.9.8 (Strong duality theorem)

Let S

⊆ Rnx be nonempty and convex. Let the functions f and gi, i = 1, . . . , ng be convex. Let

the functions h j (x) = a j x − γ j , j = 1, . . . , nh be affine linear. Let w(P ) and w(D) denote the optimal primal and dual objective function values, respectively. Let w(P ) be finite and let there exist y ∈ relint (S ) with

gi(y) < 0, i = 1, . . . , ng,

h j(y) = 0, j = 1, . . . , nh

( Slater condition). Then the dual problem is solvable and it holds w(P ) = w(D).

Proof. cf. Geiger and Kanzow [GK02], p. 323.

Remark 3.9.9 The assumptions of the strong duality theorem can be weakened further. Only

those inequalities, which are not affine, have to be satisfied strictly by some feasible y ∈ relint (S ),cf., e.g., Blum and Oettli [BO75].

The advantage of the dual problem is, that the dual objective function ψ(λ, µ) is concave on its

domain dom(ψ) := (λ, µ) ∈ R

ng+nh | λ ≥ 0ng , ψ(λ, µ) > −∞.

Hence, the problem minλ≥0ng ,µ∈Rnh −ψ(λ, µ) is a convex problem. In special cases this problemcan be solved more easily than the primal problem. Furthermore, each solution of the dualproblem is already a global maximum. If, in addition, the occurrence of a duality gap can beexcluded, then in view of Theorem 3.9.6 the primal problem can be solved indirectly by solvingthe dual problem. If a duality gap cannot be excluded, then the dual problem at least provides

lower bounds for the optimal primal objective function value, cf. Theorem 3.9.5.

3.10 Mixed-Integer Nonlinear Programs and Branch&Bound

So far, the variables x in the nonlinear optimization problems under consideration were real-valued and could assume any real number as value which belongs to a convex subset of Rn withnon-empty interior. In addition to these continuous optimization variables many optimizationproblems of practical interest include optimization variables which are restricted to a discreteset consisting of at most countably many elements, e.g. 0, 1n or Zn. These variables arecalled discrete variables. In reality, discrete variables often are used to model on/off switches,decisions, gear shifts, or discrete resource assignments.

Let setsS := x ∈Rnx | xl ≤ x ≤ xu, xl, xu ∈ R

nx, xl ≤ xu,

andY := y ∈ Z

ny | yl ≤ y ≤ yu, yl, yu ∈ Zny , yl ≤ yu

be given. Furthermore, let the functions

f : Rnx × Y → R,

g = (g1, . . . , gng) : Rnx × Y → Rng ,

h = (h1, . . . , hnh) : Rnx × Y → R

nh

be continuously differentiable w.r.t. the first n x components. We consider

c 2006 by M. Gerdts

Page 84: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 84/243

76 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

Problem 3.10.1 (Mixed Integer Nonlinear Program (MINLP))Find x ∈ Rnx and y ∈ Zny such that f (x, y) is minimized subject to the constraints

gi(x, y) ≤ 0, i = 1, . . . , ng,

h j

(x, y) = 0, j = 1, . . . , nh

,

x ∈ S,

y ∈ Y .Since Y is a finite set, the simplest idea to solve Problem 3.10.1 would be to solve Problem 3.10.1for all fixed choices y ∈ Y . For fixed y ∈ Y Problem 3.10.1 is a common differentiable nonlinearoptimization problem with optimization variables x and can be solved numerically by, e.g., SQPmethods, if a solution exists at all for the particular choice y.Just to give an idea how long this complete enumeration technique may take, consider thespecial case Y = 1, 2, 3, 4, 5ny with ny ∈ N. Depending on the dimension n y of y , the numberof admissible points in Y is given by 5ny . Assume that each numerical solution of Problem 3.10.1

with fixed vector y ∈ Y requires one second. Then, for ny = 10 this would lead to an overallsolution time of 510/(3600 · 24) ≈ 113 days. For ny = 20 it takes approximately 3 · 106 years,for ny = 40 even 3 · 1020 years. Hence, this idea of complete enumeration only works for smallcombinatorial problems.

3.10.1 Branch&Bound

The Branch&Bound algorithm basically is a tree-search algorithm combined with a rule forpruning subtrees. The root of the tree corresponds to the original problem 3.10.1. All othernodes of the tree correspond to the original problem subject to additional constraints for thediscrete variables. More precisely, the successor nodes are constructed in such a way that theunion of all feasible sets of the successors yields the feasible set of the father’s node. Hence,the optimal objective function values of the nodes are non-decreasing when traversing the treedownwards starting at the root.For each node a lower bound of the objective function of the corresponding problem has to bedetermined. Usually this is done by solving a relaxed problem or by evaluating the objectivefunction of the dual problem resp. the dual problem of the relaxed problem. If, coincidently, onefinds a feasible point for Problem 3.10.1, then the respective objective function value providesan upper bound for the optimal objective function value of Problem 3.10.1 and the node isexplored. Usually, this will not be the case and the value serves as a lower bound for the subtreeemanating from the current node. If the lower bound is less than the upper bound provided bythe best solution found so far, then all successors of the current node are generated by addingadditional constraints (branching). Afterward, all successors have to be investigated in the sameway. If at some node the lower bound is greater or equal than the upper bound provided by thebest solution found so far, then this node needs not to be explored any further and the subtreecan be pruned, since the optimal objective function values of all nodes in the subtree are greateror equal than the lower bound.For an implementable Branch&Bound algorithm the following components have to be specifiedin detail:

• algorithm for determining a lower and an upper bound of the optimal objective functionvalue of Problem 3.10.1;

• branching rule for creating new nodes;

• traversing rule for determining the order in which new nodes are generated;

c 2006 by M. Gerdts

Page 85: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 85/243

3.10. MIXED-INTEGER NONLINEAR PROGRAMS AND BRANCH&BOUND 77

• fathoming rules for pruning subtrees.

3.10.2 Bounding

Every feasible point (x, y) of Problem 3.10.1 provides an upper bound U := f (x, y) for theoptimal objective value of Problem 3.10.1.

A common way to determine a lower bound is to solve a relaxed problem and to use its objectivefunction value as a lower bound. Often there is one technical difficulty involved, namely, thefunctions f , g, and h are only defined on Rnx ×Y and not on Rnx ×Rny as it would be necessaryfor the relaxation. In the sequel, we assume that there is a way to extend the domain of f, g,and h to Rnx × Rny , e.g. by some interpolation procedure or by building convex combinations,and that the functions are continuously differentiable on this space.We concretize the term relaxation. For given vectors yl ≤ rl ≤ ru ≤ yu with rl, ru ∈ Zny define

Y (rl, ru) := y ∈ Zny | rl ≤ y ≤ ru

and

Y (rl, ru) := y ∈ R

ny

| rl ≤ y ≤ ruand consider

Problem 3.10.2 (MINLP (rl, ru))Find x ∈ Rnx and y ∈ Zny such that f (x, y) is minimized subject to the constraints

gi(x, y) ≤ 0, i = 1, . . . , ng,

h j(x, y) = 0, j = 1, . . . , nh,

x ∈ S,

y ∈ Y (rl, ru)

and the (continuous) nonlinear optimization problem

Problem 3.10.3 (N LP (rl, ru))

Find x ∈ Rnx and y ∈ Rny such that f (x, y) is minimized subject to the constraints

gi(x, y) ≤ 0, i = 1, . . . , ng,

h j(x, y) = 0, j = 1, . . . , nh,

x ∈ S,

y ∈ Y (rl, ru).

Evidently,

Y = Y (yl, yu) ⊆ Y (yl, yu), (3.10.1)

Y (rl, ru) ⊆ Y (rl, ru). (3.10.2)

The set Y (rl, ru) is called a relaxation of Y (rl, ru), because the constraint y ∈ Zny is relaxedrespectively dropped. Correspondingly, N LP (rl, ru) is called a relaxation of MINLP (rl, ru).Let f rl,ru denote the optimal objective function value of MIN LP (rl, ru) and f rl,ru the optimalobjective function value of N LP (rl, ru). As a consequence of (3.10.2) it holds

f rl,ru ≤ f rl,ru (3.10.3)

and hence, f rl,ru may serve as a lower bound for f r

l,ru.

c 2006 by M. Gerdts

Page 86: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 86/243

78 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

A second way to find a lower bound for f rl,ru is to find a feasible point of the dual problem of Problem 3.10.2. The objective function of the dual problem is given by

ψ(λ, µ) = inf x∈S, y∈Y (rl,ru)

L(x,y,λ,µ)

whereL(x,y,λ,µ) = f (x, y) + λg(x, y) + µh(x, y).

cf. Problem 3.9.3. According to the weak duality theorem 3.9.5 it holds

ψ(λ, µ) ≤ f rl,ru

for all dual feasible points λ ≥ 0ng and µ ∈ Rnh. Notice, that the evaluation of the dual objectivefunction ψ requires to solve a mixed integer nonlinear program as well. Since this problem isusually as hard as the primal problem, instead we consider the dual problem of the relaxedproblem 3.10.3. This problem possesses the objective function

ψ(λ, µ) = inf x∈S, y∈Y (rl,ru) L(x,y,λ,µ).

According to the weak duality theorem 3.9.5 and (3.10.3) it holds

ψ(λ, µ) ≤ f rl,ru ≤ f rl,ru

for all dual feasible points λ ≥ 0ng and µ ∈ Rnh. Hence, ψ(λ, µ) actually is a lower boundwhenever λ and µ are dual feasible.Evaluating ψ corresponds to solving a continuous optimization problem with ‘simple’ constraintsx ∈ S , y ∈ Y (rl, ru) only, which usually can be easier solved than the relaxed problem itself (often S imposes only box constraints).

In the sequel we concentrate on the relaxation technique. Nevertheless, all methods are appli-cable for the dual approach as well and can be adapted in a straightforward way.

3.10.3 Branching Rule

We focus on how to create new nodes. Each node corresponds to a mixed-integer nonlinearprogram MINLP (rl, ru) with

rl = (rl,1, . . . , rl,ny) ∈ Zny , ru = (ru,1, . . . , ru,ny ) ∈ Z

ny , rl = ru, yl ≤ rl ≤ ru ≤ yu.

Let (x, y) with y = (y1, . . . , yny) be an optimal solution of the relaxation N LP (rl, ru) of MINLP (rl, ru) such that y is not integral, i.e. y ∈ Y (rl, ru). Then, branching consists of thefollowing steps.

(i) Determine some index k with rl,k < yk < ru,k and maximal distance of yk to the closestinteger, i.e. k minimizes the expression |yk + 1/2 − yk|. Herein, x denotes the largestinteger not greater than x.

(ii) Create two new successor nodes MIN LP (r( j)l , r

( j)u ), j = 1, 2 with

r(1)l,i := rl,i, r

(1)u,i :=

rl,i, if i = k,yk, if i = k,

i = 1, . . . , ny,

and

r(2)l,i :=

rl,i, if i = k,

yk

+ 1, if i = k,

, r(2)u,i := ru,i, i = 1, . . . , ny.

c 2006 by M. Gerdts

Page 87: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 87/243

3.10. MIXED-INTEGER NONLINEAR PROGRAMS AND BRANCH&BOUND 79

A recursive application of the branching rule starting with MIN LP (yl, yu) as the root generatesa finite tree structure. Each edge corresponds to imposing one additional constraint to the set

Y (rl, ru). Notice, that the union of the sets Y (r( j)l , r

( j)u ), j = 1, 2 is just the set Y (rl, ru).

The branching rule is applicable in the same way, if the dual problem (of the relaxed problem) is

used to determine lower bounds. In this case, the evaluation of the dual problem yields a point

(x, y) = arg minx∈S, y∈Y (rl,ru)

L(x,y,λ,µ),

provided that such a point exists at all.

3.10.4 Fathoming Rules

A node MIN LP (rl, ru) is explored or fathomed, i.e. no further branching has to be done, if oneof the following events occur, cf. Leyffer [Ley01]:

• N LP (rl, ru) is infeasible. Hence, MINLP (rl, ru) is infeasible as well. This implies thatall nodes of the subtree are infeasible as well, since the feasible sets in the subtree are

restricted further by additional constraints.

• N LP (rl, ru) has an integer solution. Hence, this solution is optimal for MINLP (rl, ru)and feasible for MIN LP , since the feasible set of MIN LP (rl, ru) is a subset of the feasibleset of MIN LP . The corresponding optimal objective function value U serves as an upperbound for MINLP, that is U ≥ f yl,yu.

• The optimal objective function value f rl,ru of N LP (rl, ru) is greater or equal than the bestupper bound U found so far. Due to the branching rule, the optimal objective functionvalues are non-decreasing for the nodes of the subtree and consequently, the subtree canbe pruned, since no better solution exists in the subtree.

3.10.5 Traversing Rule

We use a depth-first search as the traversing rule in order to find admissible solutions for MINLPand hence upper bounds as fast as possible. In particular, that subproblem in step (ii) of the branching rule with the smallest objective function value is investigated first. Alternativestrategies are breadth-first search or some problem-specific search strategy.

3.10.6 Branch&Bound Algorithm

The overall Branch&Bound algorithm is as follows.

Algorithm 3.10.4 (Branch&Bound Algorithm)Init:

– Let U be an upper bound for the optimal objective function value of MINLP (if none is known, let U = ∞).

– Let the root of the tree MIN LP (yl, yu) be active.

while (there are active nodes) do

Select an active node MINLP (rl, ru), say, according to the traversing rule.

Solve N LP (rl, ru) (if possible) and let (x, y) be an optimal solution and f := f (x, y) the optimal objective function value of N LP (rl, ru).

if (infeasible) then Mark node as explored. endif

c 2006 by M. Gerdts

Page 88: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 88/243

80 CHAPTER 3. INFINITE AND FINITE DIMENSIONAL OPTIMIZATION PROBLEMS

if ( y ∈ Y ) then

if ( f < U ) then Save (x, y) as best solution and set U = f . endif

Mark node as explored.

endif if ( f ≥ U ) then Mark node as explored. endif

if ( f < U ) then

Apply branching rule, mark all successors as active, and mark current node as inactive.

endif

end

Since the set Y in our case is finite, the Branch&Bound algorithm will stop after finitely many

steps. Unfortunately, the number of steps may be very large depending on the problem. Hence,it may be necessary to stop the algorithm after a maximal number of nodes have been generated.An advantage of the Branch&Bound algorithm in this case is, that it provides an upper boundand a lower bound of the optimal objective function value of MINLP. The upper bound is givenby U in the algorithm. A lower bound arises, if the minimum L over the lower bounds of allactive nodes is taken. Moreover, the knowledge of an upper bound U and a lower bound L forthe optimal objective function value f yl,yu for the current tree allows to use a stopping criterionof type U −L < tol for a given tolerance tol > 0. This guarantees −tol < L − U ≤ f yl,yu −U ≤ 0and hence, the best solution found so far, which provides the upper bound U , has an objectivefunction value that lies within the tolerance tol of f yl,yu.Leyffer [Ley01] describes an algorithm, where the tree search and the NLP solution is inter-

laced. He finds that his strategy leads to improved lower bounds and is up to three times fasterthan common Branch&Bound. A different approach to solve convex MINLP without equalityconstraints is the outer approximation method depicted in the survey article of Grossmann andKravanja [GK97], cf. also Fletcher and Leyffer [FL94] and Duran and Grossmann [DG86b].According to Grossmann and Kravanja [GK97] the outer approximation method may lead tosatisfactory results even for non-convex problems if some heuristics are added. The outer ap-proximation method works iteratively. In each iteration a mixed integer linear optimizationproblem (master problem), which is a linearization of MINLP at the current iterate, has to besolved in order to obtain a new iterate for the integer variables. Then, a continuous nonlin-ear programming problem with fixed integer variables has to be solved. The master problemsyield a non-decreasing sequence of lower bounds, while the continuous programming problems

yield upper bounds. The iteration is stopped, if the upper and lower bounds are within a giventolerance.The determination of a global minimum of a nonconvex optimization problem is a difficult task.For the Branch&Bound-Algorithm it is not necessary to find the global optimum exactly butinstead it is sufficient to find a lower bound. This can be achieved by constructing convexunderestimators, cf. Meyer and Floudas [MF05], Akrotirianakis and Floudas [AF04a, AF04b],Adjiman and Floudas [AF96, AF01], Ryoo and Sahinidis [RS96], Androulakis et al. [AMF95],Sahinidis [Sah96], and Ghildyal and Sahinidis [GS01].

c 2006 by M. Gerdts

Page 89: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 89/243

Chapter 4

Local Minimum Principles

Optimal control problems subject to ordinary differential equations have a wide range of ap-plications in different disciplines like engineering sciences, chemical engineering, and economics.Necessary conditions known as ‘Maximum principles’ or ‘Minimum principles’ have been in-vestigated intensively since the 1950’s. Early proofs of the maximum principle are given byPontryagin [PBGM64] and Hestenes [Hes66]. Necessary conditions with pure state constraintsare discussed in, e.g., Jacobsen et al. [JLS71], Girsanov [Gir72], Knobloch [Kno75], Mau-rer [Mau77, Mau79], Ioffe and Tihomirov [IT79], and Kreindler [Kre82]. Neustadt [Neu76] andZeidan [Zei94] discuss optimal control problems with mixed control-state constraints. Hartl et

al. [HSV95] provide a survey on maximum principles for optimal control problems with state con-straints including an extensive list of references. Necessary conditions for variational problems,i.e. smooth optimal control problems, are developed in Bryson and Ho [BH75]. Second ordernecessary conditions and sufficient conditions are stated in Zeidan [Zei94]. Sufficient conditionsare also presented in Maurer [Mau81], Malanowski [Mal97], Maurer and Pickenhain [MP95b],and Malanowski et al. [MMP04]. Necessary conditions for optimal control problems subjectto index-1 DAE systems without state constraints and without mixed control-state constraintscan be found in de Pinho and Vinter [dPV97]. Implicit control systems are discussed in Dev-dariani and Ledyaev [DL99]. Recently, Backes [Bac06] derived necessary conditions for optimalcontrol problems subject to nonlinear quasi-linear DAEs without control and state constraints.Necessary and sufficient conditions for linear-quadratic DAE optimal control problems can be

found in Mehrmann [Meh91], Kunkel and Mehrmann [KM97], Kurina and Marz [KM04], andBackes [Bac06]. Roubicek and Valasek [RV02] discuss optimal control problems subject to Hes-senberg DAEs up to index three and obtain results that are closely related to our results. Theirtechnique for proving the results was based on an index reduction. Semi-explicit Index-1 systemsoften occur in process system engineering, cf. Hinsberger [Hin97], but also in vehicle simulation,cf. Gerdts [Ger03b], and many other fields of applications. A very important subclass of in-dex two DAEs is the stabilized descriptor form describing the motion of mechanical multi-bodysystems.

Necessary conditions are not only interesting from a theoretical point of view, but also providethe basis of the so-called indirect approach for solving optimal control problems numerically. Inthis approach the minimum principle is exploited and usually leads to a multi-point boundary

value problem, which is solved numerically by, e.g., the multiple shooting method. Nevertheless,even for the direct approach, which is based on a suitable discretization of the optimal controlproblem, the minimum principle is very important for the post-optimal approximation of ad-

joints. In this context, the multipliers resulting from the formulation of the necessary Fritz-Johnconditions for the finite dimensional discretized optimal control problem have to be related tothe multipliers of the original infinite dimensional optimal control problem in an appropriateway. It is evident that this requires the knowledge of necessary conditions for the optimal controlproblem.

In the sequel necessary conditions in terms of local minimum principles are derived for optimalcontrol problems subject to index-1 and index-2 DAEs, pure state constraints and mixed control-state constraints. The local minimum principles are based on necessary optimality conditions

81

Page 90: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 90/243

82 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

for infinite optimization problems. The special structure of the optimal control problems underconsideration is exploited and allows to obtain more regular representations for the multipliersinvolved. An additional Mangasarian-Fromowitz like constraint qualification for the optimalcontrol problem ensures the regularity of a local minimum.

Although, index-1 DAEs are easier to handle, we will start with the index-2 case and present acomplete proof. The proof for the index-1 case works similarly and will be omitted here. It canbe found in Gerdts [Ger05a].

4.1 Local Minimum Principles for Index-2 Problems

In this section we investigate the special case of Problem 1.1, where the algebraic constraint gdoes not depend on y and u. Again, let [t0, tf ] ⊂ R be a non-empty and bounded interval withfixed time points t0 < tf and U ⊆ Rnu a closed and convex set with non-empty interior. Let

ϕ : Rnx × Rnx → R,

f 0 : [t0, tf ]

×R

nx

×R

ny

×R

nu

→R,

f : [t0, tf ] ×Rnx ×Rny ×Rnu → Rnx ,

g : [t0, tf ] ×Rnx → R

ny ,

ψ : Rnx × Rnx → R

nψ ,

c : [t0, tf ] ×Rnx ×R

ny ×Rnu → R

nc ,

s : [t0, tf ] ×Rnx → R

ns

be mappings. We consider

Problem 4.1.1 (Higher Index DAE Optimal Control Problem)Find a state variable x

∈ W 1,∞([t0, tf ],R

nx), an algebraic variable y

∈ L∞([t0, tf ],R

ny), and a control variable u ∈ L∞([t0, tf ],Rnu) such that the objective function

F (x,y ,u) := ϕ(x(t0), x(tf )) +

tf

t0

f 0(t, x(t), y(t), u(t))dt (4.1.1)

is minimized subject to the higher index semi-explicit differential algebraic equation (DAE)

x(t) = f (t, x(t), y(t), u(t)) a.e. in [t0, tf ], (4.1.2)

0ny = g(t, x(t)) in [t0, tf ], (4.1.3)

the boundary conditions

ψ(x(t0), x(tf )) = 0nψ, (4.1.4)

the mixed control-state constraints

c(t, x(t), y(t), u(t)) ≤ 0nc a.e. in [t0, tf ], (4.1.5)

the pure state constraints

s(t, x(t)) ≤ 0ns in [t0, tf ], (4.1.6)

and the set constraints

u(t)

∈ U a.e. in [t0, tf ]. (4.1.7)

c 2006 by M. Gerdts

Page 91: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 91/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 83

Some definitions and terminologies are in order. (x,y,u) ∈ W 1,∞([t0, tf ],Rnx)×L∞([t0, tf ],R

ny)×L∞([t0, tf ],R

nu) is called admissible or feasible for the optimal control problem 4.1.1, if the con-straints (4.1.2)-(4.1.7) are fulfilled. An admissible pair

(x, y, u) ∈

W 1,∞([t0, tf ],Rnx)

×L∞([t0, tf ],R

ny)×

L∞([t0, tf ],Rnu) (4.1.8)

is called a weak local minimum of Problem 4.1.1, if there exists ε > 0 such that

F (x, y, u) ≤ F (x,y,u) (4.1.9)

holds for all admissible (x,y,u) with x − x1,∞ < ε, y − y∞ < ε, and u − u∞ < ε. Anadmissible pair (4.1.8) is called strong local minimum of Problem 4.1.1, if there exists ε > 0 suchthat (4.1.9) holds for all admissible (x,y,u) with x − x∞ < ε.

Notice, that strong local minima are also weak local minima. The converse is not true. Stronglocal minima are minimal w.r.t. a larger class of algebraic variables and controls. Weak localminima are only optimal w.r.t. all algebraic variables and controls in a L∞-neighborhood.

For notational convenience throughout this chapter we will use the abbreviations

ϕx0 := ϕx0(x(t0), x(tf )), f x[t] := f x(t, x(t), y(t), u(t)),

and in a similar way ϕxf

, f 0,x[t], f 0,y [t], f 0,u[t], cx[t], cy[t], cu[t], sx[t], f y[t], f u[t], gx[t], ψx0 , ψ

xf for

the respective derivatives.

Remark 4.1.2

• Since the algebraic component y is missing in (4.1.3), the DAE (4.1.2)-(4.1.3) is not an index-1 DAE. In contrast to the index-1 case in (4.2.3) the algebraic constraint (4.1.3)

cannot be solved for y and hence it has to be differentiated at least once w.r.t. time in order to obtain additional information about y. Consequently, the index of (4.1.2)-(4.1.3)is at least two and the DAE is called ‘higher index DAE’. From a formal point of view, ycan be considered as an additional control.

• Using the technique in Remark 1.8 we may assume that the control u does not appear in the algebraic constraint (4.1.3). This assumption is essential for the subsequent analysis.Without this simplification additional smoothness assumptions for the control have to be imposed.

So far, DAEs (4.1.2)-(4.1.3) without any additional properties of the functions f and g are notwell investigated and are extremely difficult to handle in view of existence and representation of

solutions. In the sequel we will exclusively consider index-2 DAEs which are characterized bythe following assumption, cf. Definition 1.5 and the remarks thereafter and Hestenes [Hes66], p.352.

Assumption 4.1.3 (Index-2 Assumption)Let the matrix

M (t) := g x[t] · f y[t]

be non-singular a.e. in [t0, tf ] and let M −1 be essentially bounded in [t0, tf ].

The optimal control problem 4.1.1 is to be reformulated as an infinite optimization problem inappropriate Banach spaces and necessary conditions for a weak local minimum are derived.

c 2006 by M. Gerdts

Page 92: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 92/243

84 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

The variables (x,y,u) are taken as elements from the Banach space

X := W 1,∞([t0, tf ],Rnx) × L∞([t0, tf ],R

ny) × L∞([t0, tf ],Rnu)

endowed with the norm

(x,y,u)

X := max

x

1,∞,

y

∞,

u

. The objective function F :

X → R is given by (4.1.1). If ϕ and f 0 are continuous w.r.t. all arguments and continuouslydifferentiable w.r.t. to x, y , and u, then F is Frechet-differentiable at (x, y, u) with

F (x, y, u)(x,y,u) = ϕx0x(t0) + ϕxf

x(tf )

+

tf

t0

f 0,x[t]x(t) + f 0,y[t]y(t) + f 0,u[t]u(t)dt,

cf. Kirsch et al. [KWW78], p. 94. Now we collect the equality constraints. The space

Z := L∞([t0, tf ],Rnx) × W 1,∞([t0, tf ],R

ny) ×Rnψ

endowed with the norm (z1, z2, z3)Z := maxz1∞, z21,∞, z32 is a Banach space. Theequality constraints of the optimal control problem are given by

H (x,y,u) = ΘZ ,

where H = (H 1, H 2, H 3) : X → Z with

H 1(x,y,u) = f (·, x(·), y(·), u(·)) − x(·),

H 2(x,y,u) = g(·, x(·)),

H 3(x,y,u) = −ψ(x(t0), x(tf )).

Remark 4.1.4 Notice, that we consider the function g[·] in the space of absolutely continuous functions and not in the space of continuous functions. The reason is that we will need the surjectivity of the operator (H 1, H 2). But in the space of continuous functions this operator is not surjective, cf. Lemma 4.1.6 below.

If f , g, and ψ are continuous w.r.t. all arguments and continuously differentiable w.r.t. to x, y ,and u, then H is continuously Frechet-differentiable at (x, y, u) with H = (H 1, H 2, H 3) and

H 1(x, y, u)(x,y,u) = f x[·]x(·) + f y[·]y(·) + f u[·]u(·) − x(·),

H 2(x, y, u)(x,y,u) = gx[·]x(·),

H 3(x, y, u)(x,y,u) = −ψx0x(t0) − ψ

xf x(tf ),

cf. Kirsch et al. [KWW78], p. 95. Now we collect the inequality constraints. The space

Y := L∞([t0, tf ],Rnc) × C ([t0, tf ],R

ns)

endowed with the norm (y1, y2)Y := maxy1∞, y2∞ is a Banach space. The inequalityconstraints of the optimal control problem are given by

G(x,y,u) ∈ K,

where G = (G1, G2) : X → Y and

G1(x,y ,u) = −

c(·, x(

·), y(

·), u(

·)), G2(x,y,u) =

−s(

·, x(

·)).

c 2006 by M. Gerdts

Page 93: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 93/243

Page 94: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 94/243

86 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

(b) The initial value problem given by (4.1.10)-(4.1.11) together with the consistent initialvalue x(t0) = x0 = Πh2(t0) + Γw has a unique solution x(·) ∈ W 1,∞([t0, tf ],R

nx), for everyw ∈ Rnx−ny , every h1(·) ∈ L∞([t0, tf ],R

nx) and every h2(·) ∈ W 1,∞([t0, tf ],Rny). The

solution is given by

x(t) = Φ(t)

Πh2(t0) + Γw + t

t0

Φ−1(τ )h(τ )dτ

in [t0, tf ], (4.1.12)

y(t) = −C (t)−1 (q (t) + Q(t)x(t)) a.e. in [t0, tf ], (4.1.13)

where the fundamental system Φ(t) ∈ Rnx×nx is the unique solution of

Φ(t) = A(t)Φ(t), Φ(t0) = I nx . (4.1.14)

(c) Let a vector b ∈ Rr and matrices C 0, C f ∈ Rr×nx be given, such that

rank ((C 0Φ(t0) + C f Φ(tf ))Γ) = r

holds for the fundamental solution Φ from (b). Then, the boundary value problem givenby (4.1.10)-(4.1.11) together with the boundary condition

C 0x(t0) + C f x(tf ) = b (4.1.15)

has a solution for every b ∈Rr.

Proof.

(a) Let x(t0) = x0 be consistent, i.e. x0 satisfies 0ny = A2(t0)x0 + h2(t0). Since C (t0) =A2(t0) · B1(t0) is supposed to be non-singular, A2(t0) has full row rank. Hence, thereexists a QR decomposition

A2(t0) = P R

0

, P = (Π1, Γ) ∈ R

nx×nx ,

where R ∈ Rny×ny is non-singular, P is orthogonal, Π1 ∈ Rnx×ny is a orthonormal basisof im(A2(t0)), Γ ∈ Rnx×(nx−ny) is a orthonormal basis of im(A2(t0))⊥ = ker(A2(t0)).Every x0 ∈ Rnx can be expressed uniquely as x0 = Π1v+Γw with v ∈ Rny and w ∈ Rnx−ny .Introducing this expression into the algebraic equation (4.1.11) yields

0ny = A2(t0)(Π1v + Γw) + h2(t0) = Rv + h2(t0) ⇒ v = −R−h2(t0).

Hence, the consistent values are characterized by

x0 = Πh2(t0) + Γw, Π := −Π1R−.

(b) We differentiate the algebraic equation (4.1.11) and obtain

0ny = A2(t)x(t) + A2(t)x(t) + h2(t)

=

A2(t) + A2(t)A1(t)

x(t) + A2(t)B1(t)y(t) + h2(t) + A2(t)h1(t)

= Q(t)x(t) + C (t)y(t) + q (t).

We exploit the non-singularity of C (t) = A2(t) · B1(t) in order to solve the equation w.r.t.y and obtain

y(t) = −

C (t)−1 (q (t) + Q(t)x(t)) .

c 2006 by M. Gerdts

Page 95: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 95/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 87

Introducing this expression into (4.1.10) yields

x(t) =

A1(t) − B1(t)C (t)−1Q(t)

x(t) + h1(t) − B1(t)C (t)−1q (t) = A(t)x(t) + h(t).

Considering the representation of consistent initial values x0 in (a) we are in the situation

as in Hermes and Lasalle [HL69], p. 36, and the assertions follow likewise.

(c) Part (c) exploits the solution formulas in (a) and (b). The boundary conditions (4.1.15)are satisfied, if

b = C 0x(t0) + C f x(tf )

= C 0 (Πh2(t0) + Γw) + C f Φ(tf )

Πh2(t0) + Γw +

tf

t0

Φ−1(τ )h(τ )dτ

= (C 0 + C f Φ(tf )) Γw + (C 0 + C f Φ(tf )) Πh2(t0)

+C f Φ(tf ) tf

t0

Φ−1(τ )h(τ )dτ.

Rearranging terms and exploiting Φ(t0) = I nx yields

(C 0Φ(t0) + C f Φ(tf )) Γw = b − (C 0 + C f Φ(tf )) Πh2(t0)

− C f Φ(tf )

tf

t0

Φ−1(τ )h(τ )dτ.

This equation is solvable for every b ∈ Rr, if the matrix (C 0Φ(t0) + C f Φ(tf )) Γ is of rankr. Then, for every b ∈ Rr there exists a w such that (4.1.15) is satisfied. Application of part (b) completes the proof.

Part (b) of Lemma 4.1.6 enables us to formulate necessary conditions for the optimal controlproblem 4.1.1 if Assumption 4.1.3 is valid.

Theorem 4.1.7 (Necessary Conditions)Let the following assumptions be fulfilled for the optimal control problem 4.1.1.

(i) Let the functions ϕ, f 0,f ,ψ,c,s be continuous w.r.t. all arguments and continuously dif- ferentiable w.r.t. x, y, and u. Let g be continuously differentiable w.r.t. all arguments.

(ii) Let U ⊆ Rnu be a closed and convex set with non-empty interior.

(iii) Let (x, y, u)

∈ X be a weak local minimum of the optimal control problem.

(iv) Let Assumption 4.1.3 be valid.

Then there exist non-trivial multipliers l0 ∈ R, η∗ ∈ Y ∗, λ∗ ∈ Z ∗ with

l0 ≥ 0, (4.1.16)

η∗ ∈ K +, (4.1.17)

η∗(G(x, y, u)) = 0, (4.1.18)

l0F (x, y, u)(x − x, y − y, u − u)

−η∗(G(x, y, u)(x − x, y − y, u − u))

−λ∗(H (x, y, u)(x

− x, y

− y, u

− u))

≥ 0

∀(x,y,u)

∈ S. (4.1.19)

c 2006 by M. Gerdts

Page 96: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 96/243

88 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

Proof. We show, that all assumptions of Theorem 3.4.2 are satisfied. Observe, that K is aclosed convex cone with vertex at zero, int(K ) is non-empty, the functions F and G are Frechet-differentiable, and H is continuously Frechet-differentiable due to the smoothness assumptions.Since U is supposed to be closed and convex with non-empty interior, the set S is closed andconvex with non-empty interior. It remains to show that im(H (x, y, u)) is not a proper densesubset of Z . According to part (b) of Lemma 4.1.6 the operator (H 1(x, y, u), H 2(x, y, u)) givenby

H 1(x, y, u), H 2(x, y, u)

(x,y,u) =

f x[·]x(·) + f y[·]y(·) + f u[·]u(·) − x(·), gx[·]x(·)is continuous, linear and surjective. Thus, we can apply Theorem 2.2.8, which yields that theimage of H (x, y, u) is closed in Z and hence im(H (x, y, u)) is not a proper dense subset in Z .Hence, all assumptions of Theorem 3.4.2 are satisfied and Theorem 3.4.2 yields (4.1.16)-(4.1.19).

Notice, that the multipliers are elements of the following dual spaces:

η∗

:= (η∗

1, η∗

2) ∈ Y ∗

= (L∞

([t0, tf ],Rnc

))∗

× (C ([t0, tf ],Rns

))∗

,λ∗ := (λ∗f , λ∗g, σ) ∈ Z ∗ = (L∞([t0, tf ],R

nx))∗ × W 1,∞([t0, tf ],R

ny)∗ ×R

nψ .

Hence, by Riesz’ representation theorem 2.4.2 the functional η∗2 admits the following represen-tation

η∗2(h) =ns

i=1

tf

t0

hi(t)dµi(t) (4.1.20)

for every continuous function h ∈ C ([t0, tf ],Rns). Herein, µi, i = 1, . . . , ns, are functions of

bounded variation. To make the representations unique, we choose µi, i = 1, . . . , ns, from thespace N BV ([t0, tf ],R), i.e. the space of normalized functions of bounded variation which arecontinuous from the right in (t0, tf ) and satisfy µi(t0) = 0, i = 1, . . . , ns.

In the sequel, the special structure of the functions F , G, and H in (4.1.19) is exploited. Fur-thermore, the fact, that the variational inequality (4.1.19) holds for all (x,y,u) ∈ S , i.e. for allx ∈ W 1,∞([t0, tf ],R

nx), all y ∈ L∞([t0, tf ],Rny), and all u ∈ U ad, is used to derive three separate

variational equalities and inequalities, respectively.Evaluation of the Frechet-derivatives in (4.1.19), using (4.1.20) and setting y = y, u = u yieldsthe variational equality

0 =

l0ϕx0 + σψx0

x(t0) +

l0ϕxf

+ σψxf

x(tf )

+

tf

t0

l0f 0,x[t]x(t)dt +ns

i=1

tf

t0

si,x[t]x(t)dµi(t)

+η∗1(c

x[·]x(·)) + λ

∗f (x(·) − f

x[·]x(·)) − λ

∗g(g

x[·]x(·))

(4.1.21)

which holds for all x ∈ W 1,∞([t0, tf ],Rnx). Similarly, we find

0 =

tf

t0

l0f 0,y[t]y(t)dt + η∗1(cy[·]y(·)) − λ∗f (f y[·]y(·)) (4.1.22)

for all y ∈ L∞([t0, tf ],Rny) and tf

t0

l0f 0,u[t]u(t)dt + η∗1(cu[·]u(·)) − λ∗f (f u[·]u(·)) ≥ 0 (4.1.23)

for all u ∈

U ad

− u

.

c 2006 by M. Gerdts

Page 97: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 97/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 89

4.1.1 Representation of Multipliers

At a first glance, the necessary conditions seem to be of little practical use since the multipliersη∗1, λ∗f , and λ∗g as elements of the dual spaces of L∞ resp. W 1,∞ do not possess a useful represen-tation. However, in this section we will derive nice representations for them. Herein, properties

of Stieltjes integrals will be exploited extensively. These results about Stieltjes integrals can befound in the books of Natanson [Nat75] and Widder [Wid46].

According to Lemma 4.1.6 the initial value problem

x(t) = f x[t]x(t) + f y[t]y(t) + h1(t), x(t0) = Πh2(t0), (4.1.24)

0ny = gx[t]x(t) + h2(t) (4.1.25)

has a solution for every h1 ∈ L∞([t0, tf ],Rnx), h2 ∈ W 1,∞([t0, tf ],R

ny). Notice, that Πh2(t0)is a special consistent initial value with w = 0nx−ny . According to (4.1.12) and (4.1.13) in

Lemma 4.1.6 the solution is given by

x(t) = Φ(t)

Πh2(t0) +

t

t0

Φ−1(τ )h(τ )dτ

, (4.1.26)

y(t) = −M (t)−1 (q (t) + Q(t)x(t))

= −M (t)−1

q (t) + Q(t)Φ(t)

Πh2(t0) +

t

t0

Φ−1(τ )h(τ )dτ

, (4.1.27)

where

h(t) = h1(t) − f y[t]M (t)−1q (t), Q(t) = d

dtgx[t] + gx[t]f x[t], q (t) = h2(t) + gx[t]h1(t),

and Φ is the solution of

Φ(t) = A(t)Φ(t), Φ(t0) = I nx , A(t) = f x[t] − f y[t]M (t)−1Q(t).

Now, let h1 ∈ L

([t0, tf ],Rnx

), h2 ∈ W

1,∞

([t0, tf ],Rny

) be arbitrary. Adding equations (4.1.22)and (4.1.21) and exploiting the linearity of the functionals leads to

l0ϕx0 + σψ

x0

x(t0) +

l0ϕxf

+ σψxf

x(tf )

+

tf

t0

l0f 0,x[t]x(t) + l0f 0,y[t]y(t)dt +ns

i=1

tf

t0

si,x[t]x(t)dµi(t)

+η∗1(cx[·]x(·) + cy[·]y(·)) + λ∗f (h1(·)) + λ∗g(h2(·)) = 0. (4.1.28)

Introducing the solution formulas (4.1.26)-(4.1.27) into equation (4.1.28) and combining terms

c 2006 by M. Gerdts

Page 98: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 98/243

90 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

leads to the expression

l0ϕx0 + σψ

x0

Πh2(t0) +

l0ϕxf

+ σψxf

Φ(tf )Πh2(t0)

+ tf

t0

l0 f 0[t]Φ(t)Πh2(t0)dt +nx

i=1

tf

t0

si,x[t]Φ(t)Πh2(t0)dµi(t)

+

l0ϕxf + σψ

xf

Φ(tf )

tf

t0

Φ−1(t)h1(t)dt

l0ϕxf + σψ

xf

Φ(tf )

tf

t0

Φ−1(t)f y[t]M (t)−1q (t)dt

+

tf

t0

l0 f 0[t]Φ(t)

t

t0

Φ−1(τ )h1(τ )dτ

dt

− tf

t0

l0 f 0[t]Φ(t) t

t0

Φ−1(τ )f y[τ ]M (τ )−1q (τ )dτ dt

− tf

t0

l0f 0,y[t]M (t)−1q (t)dt

+ns

i=1

tf

t0

si,x[t]Φ(t)

t

t0

Φ−1(τ )h1(τ )dτ

dµi(t)

−ns

i=1

tf

t0

si,x[t]Φ(t)

t

t0

Φ−1(τ )f y[τ ]M (τ )−1q (τ )dτ

dµi(t)

+η∗1(cx[·]x(·) + cy[·]y(·)) + λ∗f (h1(·)) + λ∗g(h2(·)) = 0, (4.1.29)

where

f 0[t] := f 0,x[t] − f 0,y[t]M (t)−1Q(t).

Integration by parts yields

tf

t0

f 0[t]Φ(t)

t

t0

Φ−1(τ )h1(τ )dτ

dt =

tf

t0

tf

tf 0[τ ]Φ(τ )dτ

Φ−1(t)h1(t)dt

and

tf

t0

f 0[t]Φ(t)

t

t0

Φ−1(τ )f y[τ ]M (τ )−1q (τ )dτ

dt

=

tf

t0

tf

tf 0[τ ]Φ(τ )dτ

Φ−1(t)f y[t]M (t)−1q (t)dt.

The Riemann-Stieltjes integrals are to be transformed using integration by parts. Therefore, leta : [t0, tf ]

→ Rn be continuous, b : [t0, tf ]

→ Rn absolutely continuous and µ : [t0, tf ]

→ R of

c 2006 by M. Gerdts

Page 99: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 99/243

Page 100: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 100/243

92 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

Equation (4.1.30) is equivalent with

ζ h2(t0) +

tf

t0

pf (t)h1(t)dt +

tf

t0

pg(t)q (t)dt

1(c

x[·]x(·) + c

y[·]y(·)) + λ

f (h1(·)) + λ

g(h2(·)) = 0, (4.1.31)

where

ζ :=

l0ϕx0 + σψ

x0

+

l0ϕxf + σψ

xf

Φ(tf )

+

tf

t0

l0 f 0[t]Φ(t)dt +nx

i=1

tf

t0

si,x[t]Φ(t)dµi(t)

Π

=

l0ϕx0 + σψ

x0+ pf (t0)

Π, (4.1.32)

pf (t) :=

l0ϕxf + σψ

xf

Φ(tf ) +

tf

tl0 f 0[τ ]Φ(τ )dτ

+ns

i=1

tf

tsi,x[τ ]Φ(τ )dµi(τ )

Φ−1(t), (4.1.33)

pg(t) := −l0f 0,y[t]M (t)−1

l0ϕxf + σψ

xf

Φ(tf ) +

tf

tl0 f 0[τ ]Φ(τ )dτ

+

nsi=1

tf

tsi,x[τ ]Φ(τ )dµi(τ )Φ−1(t)f y[t]M (t)−1. (4.1.34)

(4.1.31) and (4.1.23) will be exploited in order to derive explicit representations of the functionalsλ∗f , λ∗g, and η∗1. This is possible if either there are no mixed control-state constraints (4.1.5)present in the optimal control problem, or if there are no set constraints (4.1.7), i.e. U = Rnu.

In case of no mixed control-state constraints we find

Corollary 4.1.8 Let the assumptions of Theorem 4.1.7 be valid and let there be no mixed control-state constraints (4.1.5) in the optimal control problem 4.1.1. Then there exist ζ ∈ Rny

and functions pf (·) ∈ BV ([t0, tf ],Rnx), pg(·) ∈ L∞([t0, tf ],R

ny) with

λ∗f (h1(·)) = − tf

t0

pf (t) + pg(t)gx[t]

h1(t)dt, (4.1.35)

λ∗g(h2(·)) = −ζ h2(t0) − tf

t0

pg(t)h2(t)dt, (4.1.36)

for every h1 ∈ L∞([t0, tf ],Rnx) and every h2 ∈ W 1,∞([t0, tf ],R

ny).

Proof. In the absence of mixed control-state constraints equation (4.1.31) is equivalent with

ζ h2(t0) + tf

t0

pf (t)h1(t)dt + tf

t0

pg(t)q (t)dt + λ∗f (h1(

·)) + λ∗g(h2(

·)) = 0, (4.1.37)

c 2006 by M. Gerdts

Page 101: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 101/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 93

where q (t) = h2(t) + gx[t]h1(t). Setting h2(·) ≡ Θ yields

λ∗f (h1(·)) = − tf

t0

pf (t) + pg(t)gx[t]

h1(t)dt. (4.1.38)

Likewise, setting h1(·) ≡ Θ yields

λ∗g(h2(·)) = −ζ h2(t0) − tf

t0

pg(t)h2(t)dt. (4.1.39)

The function pf (·) is of bounded variation and p g(·) is essentially bounded.

In the presence of mixed control-state constraints we find

Corollary 4.1.9 Let the assumptions of Theorem 4.1.7 be valid and let U = Rnu. Let

rank (cu[t]) = nc (4.1.40)

hold almost everywhere in [t0, tf ]. Furthermore, let the pseudo-inverse of c u[t]

(cu[t])+ := cu[t]

cu[t]cu[t]−1

be essentially bounded and let the matrix

M (t) := g x[t] · f y[t] − f u[t](cu[t])+cy[t]

(4.1.41)

be non-singular with essentially bounded inverse M −1 almost everywhere in [t0, tf ].Then there exist ζ ∈ Rny and functions

pf (·) ∈ BV ([t0, tf ],Rnx

), pg(·) ∈ L∞

([t0, tf ],Rny

), η ∈ L∞

([t0, tf ],Rnc

)

with

λ∗f (h1(·)) = − tf

t0

pf (t) + pg(t)gx[t]

h1(t)dt,

λ∗g(h2(·)) = −ζ h2(t0) − tf

t0

pg(t)h2(t)dt,

η∗1(k(·)) =

tf

t0

η(t)k(t)dt

for every h1 ∈ L∞

([t0, tf ],Rnx

), every h2 ∈ W 1,∞

([t0, tf ],Rny

), and every k ∈ L∞

([t0, tf ],Rnc

).Proof. The assumption U = Rnu implies U ad = L∞([t0, tf ],R

nu) and hence, inequality (4.1.23)turns into the equality tf

t0

l0f 0,u[t]u(t)dt + η∗1(cu[·]u(·)) − λ∗f (f u[·]u(·)) = 0 (4.1.42)

for all u ∈ L∞([t0, tf ],Rnu). Equation (4.1.31) for the particular choices h1(·) = f u[·]u(·) and

h2(·) ≡ Θ implies q (t) = g x[t]h1(t) and

−λ∗f (h1(

·)) =

tf

t0 pf (t) + pg(t)gx[t]h1(t)dt + η∗1(cx[

·]x(

·) + cy [

·]y(

·)), (4.1.43)

c 2006 by M. Gerdts

Page 102: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 102/243

94 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

where x and y are determined by

x(t) = f x[t]x(t) + f y[t]y(t) + f u[t]u(t), x(t0) = 0nx , (4.1.44)

0ny = gx[t]x(t). (4.1.45)

Notice, that x(t0) = 0nx is consistent with (4.1.45). Introducing (4.1.43) into (4.1.42) andexploiting the linearity of the functional η ∗1 yields tf

t0

Hu[t]u(t)dt + η∗1(k(·)) = 0 (4.1.46)

withH

u[t] := l0f 0,u[t] + pf (t) + pg(t)gx[t]

f u[t]

andk(t) := cx[t]x(t) + cy[t]y(t) + cu[t]u(t). (4.1.47)

Due to the rank assumption (4.1.40) equation (4.1.47) can be solved for u with

u(t) = (cu[t])+ k(t) − cx[t]x(t) − cy[t]y(t) , (4.1.48)

where (cu[t])+ denotes the pseudo-inverse of c u[t]. Using the relation (4.1.48), equation (4.1.46)becomes tf

t0

Hu[t](cu[t])+

k(t) − cx[t]x(t) − cy[t]y(t)

dt + η∗1(k(·)) = 0. (4.1.49)

Replacing u in (4.1.44) by (4.1.48) yields

x(t) = f x[t]x(t) + f y[t]y(t) + f u[t](cu[t])+k(t), x(t0) = 0nx , (4.1.50)

0ny = gx[t]x(t) (4.1.51)

wheref x[t] := f x[t] − f u[t](cu[t])+cx[t], f y[t] := f y[t] − f u[t](cu[t])+cy[t].

Notice, that M (t) = g x[t] · f y[t] is assumed to be non-singular with essentially bounded inverseM −1 on [t0, tf ]. Using the solution formula in Lemma 4.1.6, the solution of (4.1.50)-(4.1.51) canbe written as

x(t) = Φ(t)

t

t0

Φ−1(τ )h(τ )dτ, y(t) = − M (t)−1

q (t) + Q(t)x(t)

,

where Φ solves the initial value problem

˙Φ(t) = f

x[t]

− f

y[t] M (t)−1 Q(t) Φ(t), Φ(t

0) = I

nx

,

h and Q are given by

h(t) = f u[t]

cu[t]+

k(t) − f y[t] M (t)−1q (t), Q(t) = d

dtgx[t] + gx[t]f x[t],

and q (t) = gx[t]f u[t] (cu[t])+ k(t). Introducing q and h into the solution formula leads to therepresentations

x(t) = Φ(t)

t

t0

w(τ )k(τ )dτ, (4.1.52)

y(t) = −

ν 1(t)x(t)−

ν 2(t)k(t) (4.1.53)

c 2006 by M. Gerdts

Page 103: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 103/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 95

where

w(t) := Φ−1(t)

f u[t] − f y[t] M (t)−1gx[t]f u[t]

cu[t]+

,

ν 1(t) := M (t)−1 Q(t),

ν 2(t) := M (t)−1gx[t]f u[t] cu[t]+ .

Equations (4.1.49), (4.1.52), and (4.1.53) and integration by parts yields

−η∗1(k(·)) =

tf

t0

Hu[t](cu[t])+

I + cy[t]ν 2(t)

k(t)dt

− tf

t0

Hu[t](cu[t])+c[t]Φ(t)

t

t0

w(τ )k(τ )dτ

dt

=

tf

t0

Hu[t](cu[t])+

I + cy[t]ν 2(t)

k(t)dt

− tf

t0

tf

tH

u[τ ](cu[τ ])+c[τ ]Φ(τ )dτ

w(t)k(t)dt

where

c[t] := cx[t] − cy[t]ν 1(t) = cx[t] − cy[t] M (t)−1 Q(t).

With the definition

η(t) :=

tf

tH

u[τ ](cu[τ ])+c[τ ]Φ(τ )dτ

w(t) − H

u[t](cu[t])+

I + cy[t]ν 2(t)

we thus obtained the representation

η∗1(k(·)) =

tf

t0

η(t)k(t)dt,

where η is an element of L∞([t0, tf ],Rnc).

Introducing this representation into (4.1.31), using integration by parts and the solution formulas(4.1.26), (4.1.27) leads to

−λ∗f (h1(·)) − λ∗g(h2(·))

= ζ h2(t0) +

tf

t0

pf (t)h1(t)dt +

tf

t0

ˆ pg(t)q (t)dt

+ tf

t0

η(t)ν 3(t)Φ(t)

Πh2(t0) + t

t0

Φ−1(τ )h1(τ )dτ

dt

− tf

t0

η(t)ν 3(t)Φ(t)

t

t0

Φ−1(τ )f y[τ ]M (τ )−1q (τ )dτ

dt

=

ζ +

tf

t0

η(t)ν 3(t)Φ(t)Πdt

h2(t0) +

tf

t0

pf (t)h1(t)dt +

tf

t0

ˆ pg(t)q (t)dt

+

tf

t0

tf

tη(τ )ν 3(τ )Φ(τ )dτ

Φ−1(t)h1(t)dt

− tf

t0 tf

t

η(τ )ν 3(τ )Φ(τ )dτ Φ−1(t)f y[t]M (t)−1q (t)dt,

c 2006 by M. Gerdts

Page 104: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 104/243

Page 105: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 105/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 97

Then there exist multipliers

l0 ∈ R, λf ∈ BV ([t0, tf ],Rnx), λg ∈ L∞([t0, tf ],R

ny), µ ∈ N BV ([t0, tf ],Rns), ζ ∈ R

ny , σ ∈ Rnψ

such that the following conditions are satisfied:

(i) l0 ≥ 0, (l0, ζ , σ , λf , λg, µ) = Θ,

(ii) Adjoint equations:

λf (t) = λf (tf ) +

tf

tH

x(τ, x(τ ), y(τ ), u(τ ), λf (τ ), λg (τ ), l0)dτ

+ns

i=1

tf

tsi,x(τ, x(τ ))dµi(τ ) in [t0, tf ], (4.1.56)

0ny = Hy(t, x(t), y(t), u(t), λf (t), λg(t), l0) a.e. in [t0, tf ]. (4.1.57)

(iii) Transversality conditions:

λf (t0) = −

l0ϕx0(x(t0), x(tf )) + σψx0(x(t0), x(tf )) + ζ gx(t0, x(t0))

, (4.1.58)

λf (tf ) = l0ϕxf

(x(t0), x(tf )) + σψxf

(x(t0), x(tf )). (4.1.59)

(iv) Optimality condition: Almost everywhere in [t0, tf ] for all u ∈ U it holds

Hu(t, x(t), y(t), u(t), λf (t), λg(t), l0)(u − u(t)) ≥ 0. (4.1.60)

(v) Complementarity condition:µi is monotonically increasing on [t0, tf ] and constant on every interval (t1, t2) with t1 < t2

and si(t, x(t)) < 0 for all t ∈ (t1, t2).

Proof. Under the above assumptions, Corollary 4.1.8 (with λf = pf , λg = pg) guaranteesthe existence of ζ ∈ Rny , λf ∈ BV ([t0, tf ],R

nx), and λg ∈ L∞([t0, tf ],Rny) such that for every

h1 ∈ L∞([t0, tf ],Rnx), h2 ∈ W 1,∞([t0, tf ],R

ny ) it holds

λ∗f (h1(·)) = − tf

t0

λf (t) + λg(t)gx[t]

h1(t)dt, (4.1.61)

λ∗g(h2(·)) = −ζ h2(t0) − tf

t0

λg(t)h2(t)dt. (4.1.62)

(a) Equation (4.1.17) is equivalent withl0ϕx0 + σψ

x0 + ζ gx[t0]

x(t0) +

l0ϕxf + σψ

xf

x(tf )

+

tf

t0

l0f 0,x[t]x(t)dt +

tf

t0

λf (t) + λg(t)gx[t]

f x[t]x(t) − x(t)

dt

+

tf

t0

λg(t) d

dt

gx[t]x(t)

dt +

nsi=1

tf

t0

si,x[t]x(t)dµi(t) = 0

for all x ∈ W 1,∞([t0, tf ],Rnx). Exploitation of

d

dt gx[t]x(t) = d

dt

gx[t] x(t) + gx[t]x(t)

c 2006 by M. Gerdts

Page 106: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 106/243

98 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

yields l0ϕx0 + σψ

x0 + ζ gx[t0]

x(t0) +

l0ϕxf + σψ

xf

x(tf )

+ tf

t0

l0f 0,x[t]x(t)dt + tf

t0

λf (t) f x[t]x(t)

− x(t) dt

+

tf

t0

λg(t)

gx[t]f x[t] + d

dtgx[t]

x(t)dt +

nsi=1

tf

t0

si,x[t]x(t)dµi(t) = 0,

which can be written equivalently asl0ϕx0 + σψ

x0 + ζ gx[t0]

x(t0) +

l0ϕxf + σψ

xf

x(tf )

+

tf

t0

Hx[t]x(t)dt +

nsi=1

tf

t0

si,x[t]x(t)dµi(t) − tf

t0

λf (t)x(t)dt = 0

for all x ∈ W 1,∞([t0, tf ],Rnx). Application of the computation rules for Stieltjes integrals

yields l0ϕx0 + σψ

x0 + ζ gx[t0]

x(t0) +

l0ϕxf + σψ

xf

x(tf )

+

tf

t0

Hx[t]x(t)dt +

nsi=1

tf

t0

si,x[t]x(t)dµi(t) − tf

t0

λf (t)dx(t) = 0.

Integration by parts of the last term leads tol0ϕx0 + σψ

x0 + ζ gx[t0] + λf (t0)

x(t0)

+

l0ϕxf + σψ

xf − λf (tf )

x(tf )

+ tf

t0Hx[t]x(t)dt +

nsi=1

tf

t0si,x[t]x(t)dµi(t) + tf

t0x(t)

dλf (t) = 0

for all x ∈ W 1,∞([t0, tf ],Rnx). This is equivalent with

l0ϕx0 + σψx0 + ζ gx[t0] + λf (t0)

x(t0)

+

l0ϕxf + σψ

xf − λf (tf )

x(tf )

+

tf

t0

x(t)d

λf (t) −

tf

tH

x[τ ]dτ −ns

i=1

tf

tsi,x[τ ]dµi(τ )

= 0

for all x

∈ W 1,∞([t0, tf ],R

nx). This implies (4.1.58)-(4.1.59) and

C = λf (t) − tf

tH

x[τ ]dτ −ns

i=1

tf

tsi,x[τ ]dµi(τ )

for some constant vector C , cf. Lemma 2.8.1 and Remark 2.8.2. Evaluation of the lastequation at t = tf yields C = λf (tf ) and thus (4.1.56).

(b) Equation (4.1.18) is equivalent with tf

t0

Hy[t]y(t)dt = 0

for all y ∈

L∞([t0, tf ],Rny). This implies (4.1.57), cf. Lemma 2.8.3.

c 2006 by M. Gerdts

Page 107: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 107/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 99

(c) Introducing (4.1.35)-(4.1.36) into (4.1.19) leads to the variational inequality tf

t0

Hu[t](u(t) − u(t))dt ≥ 0

for all u ∈ U ad. This implies the optimality condition, cf. Kirsch et al. [KWW78], p. 102.

(d) According to Theorem 4.1.7, (4.1.16) it holds η ∗2 ∈ K +2 , i.e.

η∗2(z) =ns

i=1

tf

t0

zi(t)dµi(t) ≥ 0

for all z ∈ K 2 = z ∈ C ([t0, tf ],Rns) | z(t) ≥ 0ns in [t0, tf ]. This implies, that µi is

monotonically increasing, cf. Lemma 2.8.5. Finally, the condition η ∗2(s(·, x(·))) = 0, i.e.

η∗2(s(

·, x(

·))) =

ns

i=1 tf

t0

si(t, x(t))dµi(t) = 0,

together with the monotonicity of µi implies that µi is constant in intervals with si(t, x(t)) <0, cf. Lemma 2.8.6.

Likewise, Theorem 4.1.7 and Corollary 4.1.9 yield the following necessary conditions for theoptimal control problem 4.1.1. The augmented Hamilton function is defined by

H(t,x,y,u,λf , λg, η , l0) := H(t,x,y,u,λf , λg, l0) + ηc(t,x,y,u). (4.1.63)

Theorem 4.1.11 (Local Minimum Principle for Optimal Control Problems without

Set Constraints)

Let the following assumptions be fulfilled for the optimal control problem 4.1.1.

(i) Let the functions ϕ, f 0,f ,c,s,ψ be continuous w.r.t. all arguments and continuously differ-entiable w.r.t. x, y , and u. Let g be twice continuously differentiable w.r.t. all arguments.

(ii) Let (x, y, u) ∈ X be a weak local minimum of the optimal control problem.

(iii) Let rank

cu(t, x(t), y(t), u(t))

= nc

almost everywhere in [t0, tf ].

(iv) Let the pseudo-inverse of cu[t]

(cu[t])+ = cu[t]

cu[t]cu[t]−1

(4.1.64)

be essentially bounded and let the matrix

gx[t] · f y[t] − f u[t](cu[t])+cy[t]

(4.1.65)

be non-singular almost everywhere with essentially bounded inverse in [t0, tf ].

(v) Let Assumption 4.1.3 be valid.

c 2006 by M. Gerdts

Page 108: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 108/243

100 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

(vi) Let U = Rnu.

Then there exist multipliers l0 ∈ R, ζ ∈ Rny , σ ∈ Rnψ ,

λf

∈ BV ([t0, tf ],R

nx), λg

∈ L∞([t0, tf ],R

ny), η

∈ L∞([t0, tf ],R

nc), µ

∈ N BV ([t0, tf ],R

ns)

such that the following conditions are satisfied:

(i) l0 ≥ 0, (l0, ζ , σ , λf , λg, η , µ) = Θ,

(ii) Adjoint equations:

λf (t) = λf (tf ) +

tf

tH

x(τ, x(τ ), y(τ ), u(τ ), λf (τ ), λg(τ ), η(τ ), l0)dτ

+ns

i=1

tf

tsi,x(τ, x(τ ))dµi(τ ) in [t0, tf ], (4.1.66)

0ny = Hy(t, x(t), y(t), u(t), λf (t), λg(t), η(t), l0)

a.e. in [t0, tf ]. (4.1.67)

(iii) Transversality conditions:

λf (t0) = −

l0ϕx0(x(t0), x(tf )) + σψx0(x(t0), x(tf )) + ζ gx(t0, x(t0))

, (4.1.68)

λf (tf ) = l0ϕxf

(x(t0), x(tf )) + σψxf

(x(t0), x(tf )). (4.1.69)

(iv) Optimality condition: A.e. in [t0, tf ] it holds

Hu(t, x(t), y(t), u(t), λf (t), λg(t), η(t), l0) = 0nu . (4.1.70)

(v) Complementarity conditions:Almost everywhere in [t0, tf ] it holds

η(t)c(t, x(t), y(t), u(t)) = 0, η(t) ≥ 0nc .

µi is monotonically increasing on [t0, tf ] and constant on every interval (t1, t2) with t1 < t2and si(t, x(t)) < 0 for all t ∈ (t1, t2).

Proof. Corollary 4.1.9 yields the existence of the functions λf , λg, and η and provides repre-sentations of the functionals λ∗

f , λ∗g, and η∗1. Parts (a)-(c) and the second assertion of (d) followas in the proof of Theorem 4.1.10.From

η∗1 ∈ K +1 ⇔ tf

t0

η(t)z(t)dt ≥ 0 ∀z ∈ K 1

and

η∗1(c(·, x(·), y(·), u(·))) = 0 ⇔ tf

t0

η(t)c[t]dt = 0

we conclude that

η(t) ≥ 0nc , η(t)c[t] = 0, a.e. in [t0, tf ],

cf. Lemma 2.8.5.

c 2006 by M. Gerdts

Page 109: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 109/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 101

The following considerations apply to both, Theorem 4.1.10 and Theorem 4.1.11 and differ onlyin the Hamilton functions H and H, respectively. Hence, we restrict the discussion to thesituation of Theorem 4.1.10.The multiplier µ is of bounded variation. Hence, it has at most countably many jump pointsand µ can be expressed as µ = µ

a + µ

d + µ

s, where µ

a is absolutely continuous, µ

d is a jump

function, and µs is singular (continuous, non-constant, µs = 0 a.e.). Hence, the adjoint equation(4.1.56) can be written as

λf (t) = λf (tf ) +

tf

tH

x(τ, x(τ ), y(τ ), u(τ ), λf (τ ), λg(τ ), l0)dτ

+ns

i=1

tf

tsi,x(τ, x(τ ))dµi,a(τ ) +

tf

tsi,x(τ, x(τ ))dµi,d(τ )

+

tf

tsi,x(τ, x(τ ))dµi,s(τ )

for all t

∈ [t0, tf ]. Notice, that λf is continuous from the right in (t0, tf ) since µ is normalized.

Let t j, j ∈ J , be the jump points of µ. Then, at every jump point t j it holds

limε↓0

tf

tj

si,x(τ, x(τ ))dµi,d(τ ) − tf

tj−εsi,x(τ, x(τ ))dµi,d(τ )

= −si,x(t j, x(t j)) (µi,d(t j) − µi,d(t j−)) .

Since µa is absolutely continuous and µs is continuous we obtain the jump-condition

λf (t j) − λf (t j−) = −ns

i=1

si,x(t j, x(t j)) (µi(t j) − µi(t j−)) , j ∈ J .

Furthermore, since every function of bounded variation is differentiable almost everywhere, µ

and λf are differentiable almost everywhere. Exploitation of Lemma 2.4.1 proves

Corollary 4.1.12 Let the assumptions of Theorem 4.1.10 be fulfilled. Then, λ f is differentiable almost everywhere in [t0, tf ] with

λf (t) = −Hx(t, x(t), y(t), u(t), λf (t), λg(t), l0) −

nsi=1

si,x(t, x(t))µi(t). (4.1.71)

Furthermore, the jump conditions

λf (t j ) − λf (t j−) = −ns

i=1

si,x(t j , x(t j)) (µi(t j ) − µi(t j−)) (4.1.72)

hold at every point t j ∈ (t0, tf ) of discontinuity of the multiplier µ.

Notice, that µi in (4.1.71) can be replaced by the absolutely continuous component µi,a sincethe derivatives of the jump component µi,d and the singular component µi,s are zero almosteverywhere. However, λf (t) cannot be reconstructed by simple integration of λf .A special case arises, if no state constraints are present. Then, the adjoint variable λf is evenabsolutely continuous, i.e. λf ∈ W 1,∞([t0, tf ],R

nx), and the adjoint equations (4.1.56)-(4.1.57)become

λf (t) = −Hx(t, x(t), y(t), u(t), λf (t), λg(t), l0) a.e. in [t0, tf ], (4.1.73)

0ny = H

y

(t, x(t), y(t), u(t), λf (t), λg(t), l0) a.e. in [t0, tf ]. (4.1.74)

c 2006 by M. Gerdts

Page 110: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 110/243

Page 111: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 111/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 103

(ii) There exists some (x,y,u) ∈ int(S − (x, y, u)) with

H (x, y, u)(x,y,u) = ΘZ ,

G(x, y, u) + G(x, y, u)(x,y,u) ∈ int(K ).

Herein, the interior of S = W 1,∞([t0, tf ],Rnx) × L∞([t0, tf ],R

ny) × U ad is given by

int(S ) = (x,y ,u) ∈ X | ∃ε > 0 : Bε(u(t)) ⊆ U a.e. in [t0, tf ] ,

where Bε(u) denotes the open ball with radius ε around u. The interior of K 1 is given by

int(K 1) = z ∈ L∞([t0, tf ],Rnc) | ∃ε > 0 : U ε(z) ⊆ K 1

= z ∈ L∞([t0, tf ],Rnc) | ∃ε > 0 : zi(t) ≥ ε a.e. in [t0, tf ], i = 1, . . . , nc .

The interior of K 2 is given by

int(K 2) = z ∈ C ([t0, tf ],Rns) | z(t) > 0ns in [t0, tf ] .

A sufficient condition for (i) to hold is given by the following lemma, which is an immediateconsequence of part (c) of Lemma 4.1.6. It is an extension to DAEs of a lemma that has beenused by several authors (e.g. K. Malanowski) for control problems with mixed control-stateconstraints or pure state constraints.

Lemma 4.1.14 Let Assumption 4.1.3 be valid and let

rank ψx0Φ(t0) + ψ

xf Φ(tf )Γ = nψ,

where Φ is the fundamental solution of the homogeneous linear differential equation

Φ(t) = A(t)Φ(t), Φ(t0) = I nx , t ∈ [t0, tf ]

and the columns of Γ constitute an orthonormal basis of ker (g x[t0]) and

M (t) := gx[t] · f y[t],

A(t) := f x[t] − f y[t]M (t)−1Q(t),

h(t) := h1(t) − f y[t]M (t)−1q (t),

Q(t) :=

d

dt gx[t] + g

x[t] · f

x[t],

q (t) := h2(t) + gx[t]h1(t).

Then H (x, y, u) in Problem 4.1.5 is surjective.

Proof. Let h1 ∈ L∞([t0, tf ],Rnx), h2 ∈ W 1,∞([t0, tf ],R

ny), and h3 ∈ Rnψ be given. Considerthe boundary value problem

H 1(x, y, u)(x,y ,u)(t) = h1(t) a.e. in [t0, tf ],

H 2(x, y, u)(x,y ,u)(t) = h2(t) in [t0, tf ],

H

3

(x, y, u)(x,y,u) = h3.

c 2006 by M. Gerdts

Page 112: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 112/243

104 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

More explicitly, the boundary value problem becomes

−x(t) + f x[t]x(t) + f y[t]y(t) + f u[t]u(t) = h1(t) a.e. in [t0, tf ],

gx[t]x(t) = h2(t) a.e. in [t0, tf ],

ψ

x0x(t0) + ψ

xf x(tf ) = −h3.

By the rank assumption and Assumption 4.1.3 all assumptions of Lemma 4.1.6 are satisfied andthe boundary value problem is solvable. This shows the surjectivity of the mapping H (x, y, u).

Condition (ii) is satisfied if there exist x ∈ W 1,∞([t0, tf ],Rnx), y ∈ L∞([t0, tf ],R

ny), u ∈int(U ad − u), and ε > 0 satisfying

c[t] + cx[t]x(t) + cy[t]y(t) + cu[t]u(t) ≤ −ε · e a.e. in [t0, tf ], (4.1.75)

s[t] + sx[t]x(t) < 0ns in [t0, tf ], (4.1.76)

f x[t]x(t) + f y[t]y(t) + f u[t]u(t)

− x(t) = 0nx a.e. in [t0, tf ], (4.1.77)

gx[t]x(t) = 0ny in [t0, tf ], (4.1.78)

ψx0x(t0) + ψ

xf x(tf ) = 0nψ

, (4.1.79)

where e = (1, . . . , 1) ∈ Rnc .Hence, we conclude

Theorem 4.1.15 Let the assumptions of Theorems 4.1.7, 4.1.10 or 4.1.11, and Lemma 4.1.14be fulfilled. Furthermore, let there exist x ∈ W 1,∞([t0, tf ],R

nx), y ∈ L∞([t0, tf ],Rny), and

u ∈ int (U ad − u) satisfying (4.1.75)-(4.1.79). Then it holds l0 = 1 in Theorems 4.1.7 and 4.1.10 or 4.1.11, respectively.

c 2006 by M. Gerdts

Page 113: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 113/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 105

4.1.4 Example

We will apply the local minimum principle in Theorem 4.1.10 to the subsequent index-2 DAEoptimal control problem without state constraints. The task is to minimize

30 u(t)

2

dt

subject to the equations of motion of the mathematical pendulum in Gear-Gupta-Leimkuhler(GGL) formulation given by

x1(t) = x3(t) − 2x1(t)y2(t), (4.1.80)

x2(t) = x4(t) − 2x2(t)y2(t), (4.1.81)

x3(t) = − 2x1(t)y1(t) + u(t)x2(t), (4.1.82)

x4(t) = −g − 2x2(t)y1(t) − u(t)x1(t), (4.1.83)

0 = x1(t)x3(t) + x2(t)x4(t), (4.1.84)

0 = x1(t)2

+ x2(t)2

− 1, (4.1.85)

and the boundary conditions

ψ(x(0), x(3)) := (x1(0) − 1, x2(0), x3(0), x4(0), x1(3), x3(3)) = 06. (4.1.86)

Herein, g = 9.81 denotes acceleration due to gravity. The control u is not restricted, i.e. U = R.With x = (x1, x2, x3, x4), y = (y1, y2), f 0(u) = u2, and

f (x,y ,u) = (x3 − 2x1y2, x4 − 2x2y2, −2x1y1 + ux2, −g − 2x2y1 − ux1) ,

g(x) =

x1x3 + x2x4, x2

1 + x22 − 1

the problem has the structure of Problem 4.1.1. The matrix

gx(x) · f y(x,y,u) =

x3 x4 x1 x2

2x1 2x2 0 0

0 −2x1

0 −2x2

−2x1 0−2x2 0

=

−2(x21 + x2

2) −2(x1x3 + x2x4)0 −4(x2

1 + x22)

=

−2 00 −4

is non-singular in a local minimum, hence the DAE has index two and Assumption 4.1.3 is satis-fied. The remaining assumptions of Theorem 4.1.10 are satisfied as well and necessarily there ex-ist functions λf = (λf,1, λf,2, λf,3, λf,4) ∈ W 1,∞([0, 3],R4), λg = (λg,1, λg,2) ∈ L∞([0, 3],R2),and vectors ζ = (ζ 1, ζ 2) and σ = (σ1, . . . , σ6) such that the adjoint equations (4.1.73)-(4.1.74),the transversality conditions (4.1.58)-(4.1.59), and the optimality condition (4.1.60) are satisfied.The Hamilton function (4.1.55) calculates to

H(x,y,u,λf , λg, l0) = l0u2 + λf,1 (x3 − 2x1y2) + λf,2 (x4 − 2x2y2)

+λf,3 (−2x1y1 + ux2) + λf,4 (−g − 2x2y1 − ux1)

+λg,1

−gx2 − 2y1(x2

1 + x22) + x2

3 + x24 − 2y2(x1x3 + x2x4)

+λg,2 2(x1x3 + x2x4)−

4y2(x2

1

+ x2

2

) .

c 2006 by M. Gerdts

Page 114: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 114/243

106 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

In the sequel we assume l0 = 1 (actually, the Mangasarian-Fromowitz condition is satisfied).Then, the optimality condition (4.1.60) yields

0 = 2u + λf,3x2

−λf,4x1

⇒ u =

λf,4x1 − λf,3x2

2

. (4.1.87)

The transversality conditions (4.1.58)-(4.1.59) are given by

λf (0) = (−σ1 − 2ζ 2, −σ2, −σ3 − ζ 1, −σ4) , λf (3) = (σ5, 0, σ6, 0) .

The adjoint equations (4.1.73)-(4.1.74) yield

λf,1 = 2 (λf,1y2 + λf,3y1) + λf,4u

−λg,1 (−4y1x1 − 2y2x3) − λg,2 (2x3 − 8y2x1) , (4.1.88)˙λf,2 = 2 (λf,2y2 + λf,4y1) − λf,3u−λg,1 (−g − 4y1x2 − 2y2x4) − λg,2 (2x4 − 8y2x2) , (4.1.89)

λf,3 = −λf,1 − λg,1 (2x3 − 2x1y2) − 2λg,2x1, (4.1.90)

λf,4 = −λf,2 − λg,1 (2x4 − 2x2y2) − 2λg,2x2, (4.1.91)

0 = −2

λf,3x1 + λf,4x2 + λg,1(x21 + x2

2)

, (4.1.92)

0 = −2

λf,1x1 + λf,2x2 + λg,1(x1x3 + x2x4) + 2λg,2(x21 + x2

2)

. (4.1.93)

Notice, that consistent initial values for λg1(0) and λg,2(0) could be calculated from (4.1.92)-(4.1.93) and (4.1.84)-(4.1.85) by

λg,1 = −λf,3x1 − λf,4x2, λg,2 = −λf,1x1 − λf,2x2

2 .

The differential equations (4.1.80)-(4.1.85) and (4.1.88)-(4.1.93) with u replaced by (4.1.87)together with the boundary conditions (4.1.86) and λf,2(3) = 0, λf,4(3) = 0 form a two pointboundary value problem (BVP). Notice that the DAE system has index-1 constraints (4.1.92)-(4.1.93) for λg as well as index-2 constraints (4.1.84)-(4.1.85) for y .

Numerically, the BVP is solved by a single shooting method as follows. Let z = (σ1+2ζ 2, σ2, σ3+ζ 1, σ4) denote the unknown initial values of −λf and let x(t; z), y(t; z), λf (t; z), λg(t; z) denotethe solution of the initial value problem given by (4.1.80)-(4.1.85), (4.1.88)-(4.1.93) and theinitial conditions x(0) = (1, 0, 0, 0) and λf (0) =

−z. Then, the BVP is solvable, if the nonlinear

equation

G(z) := (x1(3; z), x3(3; z), λf,2(3; z), λf,4(3; z)) = (0, 0, 0, 0)

is solvable. Numerically, the nonlinear equation is solved by Newton’s method. The requiredJacobian G

z(z) is obtained by a sensitivity analysis of the initial value problem w.r.t. z. Herein,the sensitivity DAE associated with the differential equations is employed. Figures 1–3 showthe numerical solution obtained from Newton’s method. Notice, that the initial conditions in(4.1.86) and the algebraic equations (4.1.84) and (4.1.85) contain redundant information. Hence,the multipliers σ and ζ are not unique, e.g. one may set ζ = 02. In order to obtain uniquemultipliers, one could dispense with the first and third initial condition in (4.1.86), since theseare determined by (4.1.84) and (4.1.85).

c 2006 by M. Gerdts

Page 115: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 115/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 107

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

0 0.5 1 1.5 2 2.5 3

-1.2

-1

-0.8

-0.6

-0.4

-0.2

0

0 0.5 1 1.5 2 2.5 3

-4

-3

-2

-1

0

1

2

3

0 0.5 1 1.5 2 2.5 3

-3

-2.5

-2

-1.5

-1

-0.5

0

0.5

1

1.5

0 0.5 1 1.5 2 2.5 3

0

2

4

6

8

10

12

0 0.5 1 1.5 2 2.5 3

-4e-07

-3e-07

-2e-07

-1e-07

0

1e-07

2e-07

3e-07

4e-07

5e-07

6e-07

0 0.5 1 1.5 2 2.5 3

x 1

(

·

)

x 2

(

·

)

x 3

( ·

)

x 4

( ·

)

y 1

( ·

)

y 2

( ·

)

Figure 4.1: Numerical solution of BVP resulting from the minimum principle: Differentialvariable x(t) and algebraic variable y(t) for t ∈ [0, 3].

c 2006 by M. Gerdts

Page 116: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 116/243

108 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

-20

-10

0

10

20

30

40

50

0 0.5 1 1.5 2 2.5 3

-10

0

10

20

30

40

50

60

70

0 0.5 1 1.5 2 2.5 3

-20

-15

-10

-5

0

5

10

0 0.5 1 1.5 2 2.5 3

-2

0

2

4

6

8

10

12

14

0 0.5 1 1.5 2 2.5 3

0

2

4

6

8

10

12

14

16

18

20

0 0.5 1 1.5 2 2.5 3

-2

0

2

4

6

8

10

12

0 0.5 1 1.5 2 2.5 3

λ f , 1

( ·

)

λ f , 2

( ·

)

λ f ,

3 ( ·

)

λ f ,

4 ( ·

)

λ g ,

1 ( ·

)

λ g ,

2 ( ·

)

Figure 4.2: Numerical solution of BVP resulting from the minimum principle: Adjoint variablesλf (t) and λg(t) for t ∈ [0, 3].

c 2006 by M. Gerdts

Page 117: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 117/243

4.1. LOCAL MINIMUM PRINCIPLES FOR INDEX-2 PROBLEMS 109

-3

-2

-1

0

1

2

3

0 0.5 1 1.5 2 2.5 3

u ( ·

)

Figure 4.3: Numerical solution of BVP resulting from the minimum principle: Control u(t) fort ∈ [0, 3].

Finally, the output of Newton’s method is depicted.

ITER L2 NORM OF RESIDUALS FINAL L2 NORM OF THE RESIDUALS

-------------------------------- 0.3975362397848623E-10

0 0.4328974468004262E+01 EXIT PARAMETER 1

1 0.4240642504525387E+01 FINAL APPROXIMATE SOLUTION

2 0.4193829901526831E+01 0.1520606630397864E+02

3 0.4144234257066851E+01 -0.1310630040572912E+02

4 0.4112295471233999E+01 0.1962000001208976E+02

5 0.4044231539950495E+01 -0.2769958914561794E+00

6 0.3923470866048804E+01

7 0.3705978590921865E+01

8 0.3690148344331338E+01

9 0.3310555206607231E+01

10 0.2881208390804098E+01

...

22 0.4792724723868053E-01

23 0.2221098477722281E-01

24 0.6982985591502162E-04

25 0.1271838284143635E-04

26 0.5326278869236522E-0627 0.1478490489094870E-06

28 0.1478490489094870E-06

29 0.1478490489094870E-06

30 0.4656214333457540E-08

31 0.3079526214273727E-09

32 0.1222130887100730E-09

33 0.1222130887100730E-09

34 0.4512679350279913E-10

35 0.4512679350279913E-10

36 0.3975362397848623E-10

37 0.3975362397848623E-10

38 0.3975362397848623E-10

c 2006 by M. Gerdts

Page 118: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 118/243

110 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

Remark 4.1.16 Usually, one would expect that the necessary conditions hold with the Hamilton function H = l0f 0 + λf f + λg g. The following example in Backes [Bac06] shows that this is not true:Minimize

1

2x21(tf ) +

1

2 tf

0 x3(t)2

+ u(t)2

dt

subject to

x1(t) = u(t), x1(0) = a,

x2(t) = −x3(t) + u(t), x2(0) = 0,

0 = x2(t),

where a = 0. Solution:

x1(t) = a1

t

2 + tf , x2(t) = 0, x3(t) = u(t) =

a

2 + tf

.

4.2 Local Minimum Principles for Index-1 Problems

We summarize local minimum principles for the optimal control problem 1.1, i.e. for

Problem 4.2.1 (Index-1 DAE optimal control problem)Find a state variable x ∈ W 1,∞([t0, tf ],R

nx), an algebraic variable y ∈ L∞([t0, tf ],Rny), and a

control variable u ∈ L∞([t0, tf ],Rnu) such that the objective function

F (x,y ,u) := ϕ(x(t0), x(tf )) +

tf

t0

f 0(t, x(t), y(t), u(t))dt (4.2.1)

is minimized subject to the semi-explicit index-1 DAE

x(t) = f (t, x(t), y(t), u(t)) a.e. in [t0, tf ], (4.2.2)

0ny = g(t, x(t), y(t), u(t)) a.e. in [t0, tf ], (4.2.3)

the boundary conditions

ψ(x(t0), x(tf )) = 0nψ, (4.2.4)

the mixed control-state constraints

c(t, x(t), y(t), u(t)) ≤ 0nc a.e. in [t0, tf ], (4.2.5)

the pure state constraints

s(t, x(t)) ≤ 0ns in [t0, tf ], (4.2.6)

and the set constraints

u(t) ∈ U a.e. in [t0, tf ]. (4.2.7)

We assume:

Assumption 4.2.2 (Index-1 Assumption)Let gy [t] be non-singular a.e. in [t0, tf ] and let gy[t]−1 be essentially bounded.

Assumption 4.2.2 implies that the index of the DAE (4.2.2)-(4.2.3) is one.

c 2006 by M. Gerdts

Page 119: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 119/243

Page 120: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 120/243

Page 121: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 121/243

4.2. LOCAL MINIMUM PRINCIPLES FOR INDEX-1 PROBLEMS 113

(iv) Optimality conditions: It holds

Hu(t, x(t), y(t), u(t), λf (t), λg(t), η(t), l0) = 0nu (4.2.19)

a.e. in [t0, tf ].

(v) Complementarity conditions:It holds

η(t)c(t, x(t), y(t), u(t)) = 0, η(t) ≥ 0nc

almost everywhere in [t0, tf ].

µi is monotonically increasing on [t0, tf ] and constant on every interval (t1, t2) with t1 < t2and si(t, x(t)) < 0 for all t ∈ (t1, t2).

The following considerations apply to both, Theorem 4.2.4 and Theorem 4.2.5, and differ only inthe Hamilton functions H and H, respectively. Hence, we restrict the discussion to the situationof Theorem 4.2.4.

Corollary 4.2.6 Let the assumptions of Theorem 4.2.4 be fulfilled. Then, λf is differentiable almost everywhere in [t0, tf ] with

λf (t) = −Hx(t, x(t), y(t), u(t), λf (t), λg(t), l0) −

nsi=1

si,x(t, x(t))µi(t). (4.2.20)

Furthermore, the jump conditions

λf (t j ) − λf (t j−) = −ns

i=1

si,x(t j , x(t j)) (µi(t j ) − µi(t j−)) (4.2.21)

hold at every point t j ∈ (t0, tf ) of discontinuity of the multiplier µ.

Notice, that µi in (4.2.20) can be replaced by the absolutely continuous component µi,a sincethe derivatives of the jump component µi,d and the singular component µi,s are zero almosteverywhere. However, λf (t) cannot be reconstructed by simple integration of λf .A special case arises, if no state constraints are present. Then, the adjoint variable λf is evenabsolutely continuous, i.e. λf ∈ W 1,∞([t0, tf ],R

nx), and the adjoint equations (4.2.8)-(4.2.9)become

λf (t) = −Hx(t, x(t), y(t), u(t), λf (t), λg(t), l0) a.e. in [t0, tf ], (4.2.22)

0ny = Hy(t, x(t), y(t), u(t), λf (t), λg(t), l0) a.e. in [t0, tf ]. (4.2.23)

The adjoint equations (4.2.22) and (4.2.23) form a DAE system of index one for λf and λg,where λf is the differential variable and λg denotes the algebraic variable. This follows from(4.2.23), which is given by

0ny = l0

f 0,y[t]

+

f y[t]

λf (t) +

gy[t]

λg(t).

Since g y[t] is non-singular, we obtain

λg(t) = − gy[t]−1

l0

f 0,y[t]

+

f y[t]

λf (t)

.

Finally, we address the problem of regularity and state conditions which ensure that the multi-plier l0 is not zero and, without loss of generality, can be normalized to one.

c 2006 by M. Gerdts

Page 122: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 122/243

114 CHAPTER 4. LOCAL MINIMUM PRINCIPLES

Theorem 4.2.7 Let the assumptions of Theorems 4.2.4 or 4.2.5 be fulfilled and let

rank

ψx0Φ(t0) + ψ

xf Φ(tf )

= nψ,

where Φ is the fundamental solution of the homogeneous linear differential equation

Φ(t) = A(t)Φ(t), Φ(t0) = I nx , t ∈ [t0, tf ]

with A(t) := f x[t] − f y[t]

gy[t]

−1gx[t].

Furthermore, let there exist ε > 0, x ∈ W 1,∞([t0, tf ],Rnx), y ∈ L∞([t0, tf ],R

ny), and u ∈int (U ad −u) satisfying

c[t] + cx[t]x(t) + cy[t]y(t) + cu[t]u(t) ≤ −ε · e a.e. in [t0, tf ], (4.2.24)

s[t] + sx[t]x(t) < 0ns in [t0, tf ], (4.2.25)

f x[t]x(t) + f

y[t]y(t) + f

u[t]u(t) − x(t) = 0nx a.e. in [t0, tf ], (4.2.26)

gx[t]x(t) + gy [t]y(t) + gu[t]u(t) = 0ny a.e. in [t0, tf ], (4.2.27)

ψx0x(t0) + ψ

xf x(tf ) = 0nψ

, (4.2.28)

where e = (1, . . . , 1) ∈ Rnc. Then it holds l0 = 1 in Theorems 4.2.4 or 4.2.5, respectively.

c 2006 by M. Gerdts

Page 123: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 123/243

Chapter 5

Discretization Methods for ODEs and DAEs

Many discretization methods originally designed for ODEs, e.g. Runge-Kutta methods, Multi-step methods, or extrapolation methods, can be adapted in a fairly straight forward way tosolve DAEs as well. Naturally, the resulting methods are implicit methods due to the inherentlyimplicit character of DAEs and it is necessary to solve nonlinear systems of equations in eachintegration step. The nonlinear systems usually are solved by Newton’s method or well-knownvariants thereof, e.g. globalized or simplified Newton’s method or quasi-Newton methods. Toreduce the computational effort of Newton’s method for nonlinear equations, methods based ona suitable linearization are considered, e.g. linearized implicit Runge-Kutta methods.

For notational convenience, we will write down the methods for general implicit DAE initialvalue problems of type

F (t, x(t), x(t)) = 0n, x(t0) = x0, (5.1)

where F : [t0, tf ] ×Rnx ×Rnx → Rnx, t0 ≤ t ≤ tf , and x0 is consistent. It is important to pointout, that this does not imply automatically that the resulting discretizations actually convergeto the exact solution. In contrast to the ODE case, where a quite general convergence theoryexists, for DAEs the question of convergence highly depends on the structure of the system. Inorder to obtain convergence results it is necessary to assume a certain index of (5.1) or a specialstructure in (5.1), e.g. Hessenberg structure.Generally, the discretization methods can be classified as one-step and multi-step methods. Inthe sequel, we will discuss Runge-Kutta methods (one-step) and BDF methods (multi-step),both being well investigated in the literature. In addition to these universal methods, there arealso methods which are designed for special DAEs, e.g. for the simulation of mechanical multi-body systems, cf. Simeon [Sim94], or the simulation of electric circuits, cf. Gunther [Gun95].

5.1 General Discretization Theory

Let Banach spaces (X, · X ) and (Y, · Y ) and a mapping G : X → Y be given. The task isto find x ∈ X with

G(x) = ΘY .

This problem is referred to by the triple P = (X ,Y,G). It is assumed that a unique solution xexists, that is, problem P is uniquely solvable.

Definition 5.1.1 (Discretization Method, Discretization)

(i) A discretization method M for P is a sequence (X N , Y N , ∆N , δ N , ϕN )N ∈N, where

– (X N , · X N ) and (Y N , · Y N

) are finite dimensional Banach spaces with dim (X N ) =dim (Y N );

– ∆N : X → X N and δ N : Y → Y N are linear mappings with

limN →∞

∆N (x)X N = xX , (5.1.1)

limN →∞

δ N (y)Y N = yY (5.1.2)

for each fixed x∈

X and y ∈

Y ;

115

Page 124: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 124/243

116 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

– ϕN : (X → Y ) → (X N → Y N ) is a mapping with G in the domain of ϕN for all N ∈ N.

(ii) A discretization D of the problem P generated by the discretization method M is a sequence (X N , Y N , GN )N ∈N with GN = ϕN (G) : X N

→Y N .

A solution of the discretization D is a sequence (xN )N ∈N with

GN (xN ) = ΘY N , N ∈ N.

Definition 5.1.2 (Consistency)A discretization method M = (X N , Y N , ∆N , δ N , ϕN )N ∈N of the problem P is called consistentat x ∈ X , if x is in the domain of G and of ϕN (G)(∆N (·)) for N ∈ N and

limN →∞

ϕN (G)(∆N (x)) − δ N (G(x))Y N = 0.

It is called consistent of order p at x, if

ϕN (G)(∆N (x)) − δ N (G(x))Y N = O

1

N p

as N → ∞.

If M is consistent (of order p) at x, the corresponding discretization D is called consistent (of order p) at x.

Later on we will need consistency only in the exact solution x of the original problem P . Notice,that the exact solution satisfies δ N (G(x)) = ΘY N

. Thus, it suffices to investigate the localdiscretization error defined below.

Definition 5.1.3 (Local and Global Discretization Error, Convergence)

Let x denote the solution of P .(i) Let the discretization method M be consistent. The local discretization error of M and D

is defined as the sequence lN N ∈N where

lN = ϕN (G)(∆N (x)) ∈ Y N , N ∈ N.

(ii) Let the discretization D of the problem P generated by the discretization method M possess unique solutions xN , N ∈ N. The sequence eN N ∈N with

eN = xN − ∆N (x) ∈ X N , N ∈ N

is called global discretization error of M

and D

.

(iii) M and D are called convergent, if

limN →∞

eN X N = 0,

and convergent of order p, if

eN X N = O

1

N p

as N → ∞.

In view of a proof of convergence the discretization method has to be not only consistent butalso stable.

c 2006 by M. Gerdts

Page 125: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 125/243

5.1. GENERAL DISCRETIZATION THEORY 117

Definition 5.1.4 (Stability)The discretization D is called stable at (xN )N ∈N ⊂ X N , if there exist constants S and r > 0independent of N such that uniformly for almost all N ∈ N it holds the following: Whenever

x(i)N ∈ X N , i = 1, 2, satisfy

GN (x(i)N ) − GN (xN )Y N

< r, i = 1, 2

then it holds x

(1)N − x

(2)N X N

≤ S GN (x(1)N ) − GN (x

(2)N )Y N

.

S is called stability bound, r is called stability threshold.The discretization method M is called stable, if the generated discretization D is stable at (∆N (x))N ∈N, where x is the solution of the original problem P .So far, we only assumed the unique solvability of the original problem P . It is not clear, thatthe discrete problems GN (xN ) = ΘY N

are (uniquely) solvable. It will turn out that consistency,stability and a continuity argument will guarantee unique solvability. The proofs of the followingstatements can be found in Stetter [Ste73], pp. 11-15.

Lemma 5.1.5 (Stetter [Ste73], Lem. 1.2.1, p. 11)Let GN : X N → Y N be defined and continuous in

BR(xN ) = xN ∈ X N | xN − xN X N < R

for xN ∈ X N . Furthermore, for all x(i)N ∈ BR(xN ), i = 1, 2, with

GN (x(i)N ) ∈ U r(GN (xN )) = yN ∈ Y N | yN − GN (xN )Y N

< r

let x(1)N − x

(2)N X N

≤ S GN (x(1)N ) − GN (x

(2)N )Y N

for some constants r ,S > 0. Then G−1N : Y N → X N exists and is Lipschitz continuous in

U r0(GN (xN )) with Lipschitz constant S , where r0 = minr,R/S .

Exploitation of this lemma allows to proof unique solvability. It holds

Theorem 5.1.6 (Stetter [Ste73], Th. 1.2.3, p.12)

Let x denote the unique solution of P . Let the discretization method M = (X N , Y N , ∆N , δ N , ϕN )satisfy the following conditions.

(i) GN = ϕN (G) is defined and continuous in

BR(∆N (x)) = xN ∈ X N | xN − ∆N (x)X N < R

with R > 0 independent of N .

(ii) M is consistent at x.

(iii) M is stable.

Then the discretization D = (X N , Y N , GN ) possesses unique solutions xN ∈ X N for all suffi-ciently large N ∈ N.

With Theorem 5.1.6 we can proof convergence.

c 2006 by M. Gerdts

Page 126: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 126/243

118 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

Theorem 5.1.7 (Stetter [Ste73], Th. 1.2.4, p.13)Let the assumptions of Theorem 5.1.6 be valid. Then, M is convergent. Furthermore, if M is consistent of order p, then it is convergent of order p.

Proof. Theorem 5.1.6 guarantees the existence of unique solutions xN of GN (xN ) = ΘY N for

sufficiently large N . It holds

−lN = −GN (∆N (x)) = GN (xN ) =ΘY N

−GN (∆N (x)) = GN (∆N (x) + eN ) − GN (∆N (x)).

Stability at ∆N (x) implies

eN X N = xN − ∆N (x)X N

≤ S GN (xN ) − GN (∆N (x))Y N = S lN Y N

for lN Y N = GN (xN )−GN (∆N (x))Y N

< r. Since lN Y N → 0 respectively lN Y N

= O(N − p)holds, it follows eN X N

→ 0 respectively eN X N = O(N − p).

Now, we reformulate the initial value problem (5.1) as an abstract root finding problem. The

Banach spaces (X, · X ) and (Y, · Y ) given by

X = C 1([t0, tf ],Rnx),

x(·)X = maxt0≤t≤tf

x(t),

Y = Rnx × C ([t0, tf ],R

nx),

(r, s(·))Y = r + maxt0≤t≤tf

s(t).

The mapping G : X → Y is defined by

G(x) =

x(t0) − x0

F (

·, x(

·), x(

·))

.

Then, (5.1) is equivalent with the problem

G(x) = ΘY .

In order to discretize the abstract problem, we introduce a grid

GN = ti | i = 0, 1, . . . , N depending on the discretization parameter N ∈ N with distinct grid points

t0 < t1 < .. . < tN =: tf

and step lengths h j = t j+1

−t j, j = 0, 1, . . . , N

−1. It is assumed, that the step sizes linearly

tend to zero as N approaches infinity, i.e. there exist constants c1, c2 > 0 such that

c1N

≤ h j ≤ c2N

, j = 0, 1, . . . , N − 1

for all N ∈ N. Often, we will simply consider constant step lengths h = (tf − t0)/N . Finitedimensional Banach spaces (X N , · X N

) and (Y N , · Y N ) are given by

X N = xN : GN → Rnx,

xN (·)X N = max

t∈GN

xN (t),

Y N = X N ,

· Y N

= ·

X N .

c 2006 by M. Gerdts

Page 127: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 127/243

5.1. GENERAL DISCRETIZATION THEORY 119

The restriction operators ∆N : X → X N and δ N : Y → Y N are defined by

∆N (x)(t j ) := x(t j ),

δ N (r, s(

·))(t j ) :=

r, if j = 0,

s(t j−1), if 1 ≤ j ≤ N.In the sequel, we focus on one-step methods with increment function Φ defined by

xN (t0) = x0,

xN (t j ) − xN (t j−1)

h j−1= Φ(t j−1, xN (t j−1), h j−1), j = 1, 2, . . . , N .

(5.1.3)

For the one-step method (5.1.3) the mappings ϕN : (X → Y ) → (X N → Y N ) and GN = ϕN (G) :X N → Y N are given by

ϕN (G)(xN )(t j ) =

xN (t0)−

x0, if j = 0,xN (t j) − xN (t j−1)

h j−1− Φ(t j−1, xN (t j−1), h j−1), if j = 1, . . . , N .

The following stability result holds for one-step methods.

Theorem 5.1.8 Let Φ be a locally Lipschitz continuous function w.r.t. x uniformly with respect to t and h in the exact solution x for sufficiently small step sizes h. Then, the one-step method M = (X N , Y N , ∆N , δ N , ϕN )N ∈N is stable.

Proof. Since Φ is assumed to be locally Lipschitz continuous at x, there exist constants R > 0,hR > 0, and LR > 0 such that

Φ(t, x(1), h) − Φ(t, x(2), h) ≤ LRx(1) − x(2) (5.1.4)

holds for all 0 < h ≤ hR and all

(t, x(i)) ∈ K, i = 1, 2,

whereK := (t, x) ∈R×R

nx | t0 ≤ t ≤ tf , x − x(t) ≤ R.

Let GN be a grid withmax

j=0,...,N −1h j ≤ hR. (5.1.5)

Furthermore, let grid functions x(i)N ∈ X N , i = 1, 2, be given with

x(i)N − ∆N (x)X N

≤ R i = 1, 2. (5.1.6)

Notice, that (t j , x(i)N (t j)) ∈ K for all t j ∈ GN , i = 1, 2. Define

ε(i)N := GN (x

(i)N ), i = 1, 2.

Then, for j = 0 we have

x(1)

N

(t0)−

x(2)

N

(t0) = x0 + ε(1)

N

(t0)−

(x0 + ε(2)

N

(t0)) = ε(1)

N

(t0)−

ε(2)

N

(t0).

c 2006 by M. Gerdts

Page 128: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 128/243

120 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

For j = 1, . . . , N it holds

x(1)N (t j) − x

(2)N (t j ) = x

(1)N (t j−1) + h j−1Φ(t j−1, x

(1)N (t j−1), h j−1) + h j−1ε

(1)N (t j )

− x(2)N (t j−1) − h j−1Φ(t j−1, x

(2)N (t j−1), h j−1) − h j−1ε

(2)N (t j)

≤ (1 + h j−1LR)x(1)N (t j−1) − x(2)N (t j−1) + h j−1ε(1)N (t j) − ε(2)N (t j).

Recursive evaluation leads to

x(1)N (t j ) − x

(2)N (t j) ≤

j−1k=0

(1 + hkLR)

x

(1)N (t0) − x

(2)N (t0)

+

jk=1

hk−1

j−1l=k

(1 + hlLR)

ε

(1)N (tk) − ε

(2)N (tk)

≤ exp((t j − t0)LR)ε(1)N (t0) − ε

(2)N (t0)

+ maxk=1,...,j ε

(1)

N (tk) − ε

(2)

N (tk) exp((t j − t0)LR)(t j − t0)

≤ C exp((t j − t0)LR) maxk=0,...,j

ε(1)N (tk) − ε

(2)N (tk), (5.1.7)

where C = maxtf − t0, 1. Herein, we exploited

(1 + hkLR) ≤ exp(hkLR),

j−1k=0

hk = t j − t0.

Finally, we found the estimate

x(1)

N −x(2)

N X

N ≤ S

ε(1)

N −ε(2)

N Y N

= S

GN (x(1)

N )

−GN (x

(2)

N )

Y N

(5.1.8)

with S = C exp((tf − t0)LR), which holds for all grids with (5.1.5) and all x(i)N ∈ X N , i = 1, 2,

with (5.1.6).

To complete the proof it remains to show that (5.1.8) holds, whenever x(i)N ∈ X N , i = 1, 2, satisfy

GN (x(i)N ) − GN (∆N (x))Y N

≤ r, i = 1, 2, (5.1.9)

for some r > 0. So, let (5.1.9) be valid with r ≤ RS . Then, the estimate (5.1.7) applied to

x(1)N = ∆N (x) and x

(2)N = ∆N (x), respectively, yields x

(i)N (t j)− x(t j) ≤ R for all j = 0, 1, . . . , N

and i = 1, 2. Hence, (5.1.6) and thus (5.1.9) hold. This shows the stability of the one-stepmethod with stability bound S and stability threshold R/S .

5.2 Backward Differentiation Formulae (BDF)

The Backward Differentiation Formulae (BDF) were introduced by Curtiss and Hirschfelder[CH52] and Gear [Gea71] and belong to the class of implicit linear multi-step methods.We consider the integration step from tm+k−1 to tm+k. Given approximations xN (tm), xN (tm+1),. . ., xN (tm+k−1) for x(tm), x(tm+1), . . . , x(tm+k−1) and the approximation xN (tm+k) for x(tm+k),which has to be determined, we compute the interpolating polynomial Q(t) of degree k withQ(tm+ j ) = xN (tm+ j), j = 0, . . . , k. The unknown value xN (tm+k) is determined by the postu-lation that the interpolating polynomial Q satisfies the DAE (5.1) at the time point tm+k, i.e.xN (tm+k) satisfies

F (tm+k, Q(tm+k), Q(tm+k)) = 0nx . (5.2.1)

c 2006 by M. Gerdts

Page 129: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 129/243

5.3. IMPLICIT RUNGE-KUTTA METHODS 121

The derivative Q(tm+k) can be expressed as

Q(tm+k) = 1

hm+k−1

k j=0

α jxN (tm+ j ), (5.2.2)

where the coefficients α j depend on the step sizes hm+ j−1 = tm+ j − tm+ j−1, j = 1, . . . , k.Equation (5.2.1) together with (5.2.2) yields

F

tm+k, xN (tm+k),

1

hm+k−1

k j=0

α jxN (tm+ j )

= 0nx , (5.2.3)

which is a nonlinear equation for x = xN (tm+k). Numerically, equation (5.2.3) is solved by the(globalized) Newton method given by

F x + αk

hm+k−1

F x∆[l] =

−F,

x[l+1] = x[l] + λl∆[l], l = 0, 1, . . . ,

(5.2.4)

where the functions are evaluated at (tm+k, x[l], (αkx[l] +k−1

j=0 α j xN (tm+ j))/hm+k−1).BDF methods are appealing since they only require to solve a n x-dimensional nonlinear systeminstead of a nx · s dimensional system that arises for Runge-Kutta methods.Brenan et al. [BCP96] develop in their code Dassl an efficient algorithm to set up the inter-polating polynomial by use of a modified version of Newton’s divided differences. In addition,a step size and order control based on an estimation of the local discretization error is imple-mented. A further improvement of the numerical performance is achieved by using some sort of simplified Newton’s method, where the iteration matrix F x + αk/hm+k−1

·F x is held constant

as long as possible, even for several integration steps. For higher index systems the estimationprocedure for the local discretization error has to be modified, cf. Petzold [Pet82b]. In practice,the higher index algebraic components are assigned very large error tolerances or are simplyneglected in the error estimator. Fuhrer [Fuh88] and Fuhrer and Leimkuhler [FL91] apply BDFto overdetermined DAEs. The implementation is called Odassl and is an extension of Dassl.Gear and Petzold [GP84] and Petzold [Pet89] prove the convergence for BDF methods up to order6, when applied to index-1 systems (5.1) with constant step size. Lotstedt and Petzold [LP86]show convergence up to order 6 for constant step size and certain DAEs arising in electric circuitsimulation, fluid mechanics, and mechanics. Brenan and Engquist [BE88] prove convergence forDAEs of Hessenberg type (1.14) with index 2 and 3 at constant step size. Gear et al. [GLG85]extend the results in Gear and Petzold [GP84] to BDF methods with non-constant step size and

non-constant order for index-2 DAEs (1.14) and index-1 DAEs (1.12)-(1.13). They also show,that BDF with non-constant step sizes applied to index-3 DAEs of Hessenberg type may yieldwrong results in certain components of x. All these results assume consistent initial values.

5.3 Implicit Runge-Kutta Methods

An implicit Runge-Kutta (IRK) method with s stages is defined by the Butcher array

c1 a11 · · · a1s...

... . . .

...cs as1 · · · ass

b1 · · ·

bs

(5.3.1)

c 2006 by M. Gerdts

Page 130: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 130/243

122 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

with real valued entries. An integration step of the IRK method from t m to tm+1 with step sizehm = tm+1 − tm starting at an approximation x N (tm) of x(tm) is defined by

xN (tm+1) = xN (tm) + hm

s

j=1

b j k j (tm, xN (tm), hm). (5.3.2)

It is a one-step method with increment function

Φ(tm, xN (tm), hm) :=s

j=1

b j k j (tm, xN (tm), hm). (5.3.3)

Herein, the stage derivatives k j = k j(tm, xN (tm), hm), j = 1, . . . , s, are implicitly defined by thenonlinear equation

G(k, tm, xN (tm), hm) := F

tm + c1hm, x

(1)m+1, k1

...

F

tm + cshm, x(s)m+1, ks

= 0s·nx (5.3.4)

for k = (k1, . . . , ks) ∈ Rs·nx. The quantities

x(i)m+1 = xN (tm) + hm

s j=1

aij k j (tm, xN (tm), hm), i = 1, . . . , s , (5.3.5)

can be interpreted as approximations at the intermediate time points (stages) tm + cihm.Numerically, (5.3.4) is solved by the (globalized) Newton method defined by

Gk(k[l], tm, xN (tm), hm) ∆[l] = −G(k[l], tm, xN (tm), hm),

k[l+1] = k[l] + λl∆[l], l = 0, 1, . . . .

(5.3.6)

The derivative G k evaluated at (k[l], tm, xN (tm), hm) is given by

Gk =

F x(ξ

[l]1 )

. . .

F x(ξ [l]s )

+ hm

a11F x(ξ

[l]1 ) · · · a1sF x(ξ

[l]1 )

... . . .

...

as1F x(ξ [l]s ) · · · assF x(ξ

[l]s )

, (5.3.7)

whereξ [l]i =

tm + cihm, xN (tm) + hm

s j=1

aij k[l]

j , k[l]i

, i = 1, . . . , s .

A Runge-Kutta method is called stiffly accurate, if it satisfies the conditions

cs = 1 and asj = b j, j = 1, . . . , s ,

cf. Strehmel and Weiner [SW95] and Hairer et al. [HLR89]. For instance, the RADAUIIAmethod is stiffly accurate. These methods are particularly well-suited for DAEs since the stage

x(s)m+1 and the new approximation xN (tm+1) coincide. Since x

(s)m+1 satisfies the DAE at tm + h

so does xN (tm+1).

c 2006 by M. Gerdts

Page 131: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 131/243

5.4. LINEARIZED IMPLICIT RUNGE-KUTTA METHODS 123

Given a Runge-Kutta method that is not stiffly accurate, the approximation (5.3.2) usually will

not satisfy the DAE (5.1), although the stages x(i)m+1, i = 1, . . . , s satisfy it. To circumvent this

problem, Ascher and Petzold [AP91] discuss projected Runge-Kutta methods for semi-explicitDAEs. Herein, xN (tm+1) is projected onto the algebraic equations and the projected value is

taken as approximation at tm+1.The dimension of the equation (5.3.6) can be reduced for half-explicit Runge-Kutta methods, cf.Arnold [Arn98] and Arnold and Murua [AM98]. Pytlak [Pyt98] uses IRK methods in connectionwith implicit index-1 DAEs of type F (x,x,y) = 0. Brenan et al. [BCP96] show convergenceof IRK methods for index-1 systems (5.1). Brenan and Petzold [BP89] prove convergence of IRK methods for semi-explicit index-2 DAEs. Jay [Jay93, Jay95] proves convergence for semi-explicit Hessenberg index-3 DAEs. Consistent initial values and certain regularity and stabilityconditions of the Runge-Kutta methods under consideration are assumed, cf. also Hairer etal. [HLR89].

5.4 Linearized Implicit Runge-Kutta Methods

The nonlinear system (5.3.4) has to be solved in each integration step. In addition, the DAE(5.1) itself has to be solved many times during the iterative solution of discretized optimal controlproblems. Since large step sizes h may cause Newton’s method to fail the implicit Runge-Kuttamethod usually is extended by an algorithm for automatic step size selection. The resultingmethod may be quite time consuming in practical computations. To reduce the computationalcost for integration of (5.1) we introduce linearized implicit Runge-Kutta (LRK) methods, whichonly require to solve linear equations in each integration step and allow to use fixed step-sizesduring integration. Numerical experiments show that these methods often lead to a speed-up inthe numerical solution (with a reasonable accuracy) when compared to nonlinear implicit Runge-Kutta methods or BDF methods with step-size and order selection discussed in Section 5.2.Consider the implicit Runge-Kutta method (5.3.2) with coefficients (5.3.1) and stage derivatives

implicitly defined by (5.3.4). The linearized implicit Runge-Kutta method is based on the ideato perform only one iteration of Newton’s method for (5.3.4). This leads to

Gk(k[0], tm, xN (tm), hm) ·

k − k[0]

= −G(k[0], tm, xN (tm), hm), (5.4.1)

which is a linear equation for the stage derivatives k = (k1, . . . , ks). The derivative Gk is given

by (5.3.7) and the new approximation xN (tm+1) by (5.3.2).

Remark 5.4.1 A similar idea is used for the construction of Rosenbrock-Wanner (ROW) meth-ods, compare [HW96].

The vector k [0] = (k[0]1 , . . . , k

[0]s ) denotes an initial guess for k and will be specified in the sequel.

It turns out that the choice of k

[0]

is important in view of the order of convergence.To gain more insights into the convergence properties of the linearized Runge-Kutta method thediscussion is restricted to explicit ordinary differential equations with

F (t, x(t), x(t)) := x(t) − f (t, x(t)) = 0nx , x(t0) = x0. (5.4.2)

Later on, an extension to DAEs is suggested which yields promising numerical results althougha convergence proof could not be obtained yet.For the explicit ODE in (5.4.2) the derivative G

k becomes

Gk(k[0], tm, xN (tm), hm) =

I

. . .

I

− hm

a11f x(ν 1) · · · a1sf x(ν 1)

... . . .

...a

s1f

x(ν

s) · · ·

ass

f x

(ν s)

, (5.4.3)

c 2006 by M. Gerdts

Page 132: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 132/243

124 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

where

ν i =

tm + cihm, xN (tm) + hm

sl=1

ailk[0]l

, i = 1, . . . , s .

Writing down the linear equation (5.4.1) in its components, we obtain

ki = f (ν i) + hm

s j=1

aij f x(ν i)

k j − k[0]

j

, i = 1, . . . , s . (5.4.4)

We are up to investigate the convergence properties of the LRK method for two different choicesof the initial guess k[0]. We intend to apply Theorem 5.1.7. Hence, to show convergence wehave to prove consistency and stability both in the exact solution x of the initial value problem(5.4.2). The order of consistency is determined by Taylor expansion of the local discretizationerror lN . The multivariate Taylor formula for a p + 1 times continuously differentiable functiong : Rn → Rm is given by

g(x + h) = p

j=0

1

j!g( j)(x) · (h , . . . , h)

j−fold

+O(h p+1)

for any x ∈ Rn and sufficiently small h ∈ Rn. Herein, the jth derivative of g is a j-linear mappingwith

g( j)(x) · (h1, . . . , h j ) =n

i1,...,ij=1

∂ jg(x)

∂xi1 · · · ∂xij

h1,i1 · · · h j,ij .

For notational convenience we use the abbreviations h = hm and xm = xN (tm).

5.4.1 A first choice: Constant Prediction

The simplest idea one might have is to use k [0] = (k[0]1 , . . . , k

[0]s ) with k

[0]i = 0nx for all i = 1, . . . , s,

which means, that we use xm as predictor for x(i)m+1, i = 1, . . . , s, in (5.3.5). Equation (5.4.4)

reduces to

ki = f (tm + cih, xm) + hs

j=1

aijf x(tm + cih, xm)(k j ), i = 1, . . . , s . (5.4.5)

In matrix notation, Equation (5.4.5) is given by

(I − hB1(tm, xm, h)) k = c1(tm, xm, h), (5.4.6)

where

B1(tm, xm, h) :=

a1 ⊗ f x(tm + c1h, xm)

...as ⊗ f x(tm + csh, xm)

,

c1(tm, xm, h) :=

f (tm + c1h, xm)

...f (tm + csh, xm)

,

(5.4.7)

and ai denotes the ith row of A = (aij ) and ⊗

is the Kronecker product.

c 2006 by M. Gerdts

Page 133: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 133/243

5.4. LINEARIZED IMPLICIT RUNGE-KUTTA METHODS 125

Taylor expansion of ki at the exact solution x and recursive evaluation leads to

ki = f + h

cif t +

s j=1

aijf x(k j )

+ h2

c2i2

f tt + ci

s j=1

aijf xt(k j )

+ O(h3)

= f + h

cif t +s

j=1

aijf x

f + h

c j f t +

sl=1

a jl f x(kl)

+ O(h2)

+ h2

c2

i

2 f tt + ci

s j=1

aij f xt (f + O(h))

+ O(h3)

= f + h

cif t +

s j=1

aijf x(f )

+ h2 s

j=1aij c j f

x(f

t) +

s j=1

aij

sl=1

a jl f x(f

x(f )) +

c2i

2 f tt + ci

s j=1

aijf xt(f )

+ O(h3),

where f is evaluated at (tm, x(tm)). Taylor expansion of the exact solution yields

x(tm + h) − x(tm)

h = x(tm) +

h

2!xtm +

h2

3!

...x (tm) + O(h3)

= f + h

2!

f t + f x(f )

+

h2

3! f tt + 2f tx(f ) + f x(f t) + f x(f x(f )) + f xx(f, f ) + O(h3).

Hence, the Taylor expansion of the local discretization error is given by

lN = x(tm + h) − x(tm)

h −

si=1

biki

=

1 −

si=1

bi

f + h

1

2 −

si=1

bici

f t +

1

2 −

si=1

bi

s j=1

aij

f x(f )

+ O(h2).

Notice, that the derivative f yy does not occur in the Taylor expansion of the numerical solution.Hence, the maximum attainable order is 2 and it holds

Theorem 5.4.2 Let f possess continuous and bounded partial derivatives up to order two. The LRK method (5.4.5) is consistent of order 1, if

si=1

bi = 1

holds. It is consistent of order 2, if in addition

si=1

bici = 1

2,

si=1

bi

s j=1

aij = 1

2

hold. p = 2 is the maximum attainable order of consistency.

c 2006 by M. Gerdts

Page 134: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 134/243

126 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

The LRK method (5.4.5) has a drawback. The method is not invariant under autonomization,i.e. it does not produce the same numerical results for the system (5.4.2) and the equivalentautonomous system

y(t)z(t) = f (z(t), y(t))

1 , y(0)z(0) = x0

t0 , (5.4.8)

cf. Deuflhard and Bornemann [DB02], pp. 137-138. For, discretization of (5.4.2) and (5.4.8)with the LRK method (5.4.1) yields

ki = f (tm + cih, xm) + hs

j=1

aijf x(tm + cih, xm)(k j ), i = 1, . . . , s ,

and

kyi

kzi

= f (zm, ym) + h

s j=1

aij f x(zm, ym)(k

y j ) + f

t(zm, ym)(k

z j )

1

, i = 1, . . . , s ,

respectively. Obviously, ki and k yi are different if f explicitly depends on t.

5.4.2 A second choice: Linear Prediction

We use the initial guess k[0] = (k[0]1 , . . . , k

[0]s ) with k

[0]i = f (tm, xm) for all i = 1, . . . , s, which

means, that we use xm + hs

j=1 aijf (tm, xm) as predictor for x(i)m+1, i = 1, . . . , s, in (5.3.5).

Equation (5.4.4) reduces to

ki = f (ν i) + hs

j=1

aijf x(ν i)(k j − f (tm, xm)), i = 1, . . . , s , (5.4.9)

where

ν i =

tm + cih, xm + h

sl=1

ailf (tm, xm)

, i = 1, . . . , s .

It turns out that the method (5.4.9) is invariant under autonomization under appropriate as-sumptions, which are identical to those in Deuflhard and Bornemann [DB02], Lemma 4.16, p.138.

Lemma 5.4.3 Let the coefficients (5.3.1) fulfill the conditions

si=1

bi = 1

and

ci =s

j=1

aij, i = 1, . . . , s . (5.4.10)

Then, the LRK method (5.4.9) is invariant under autonomization.

c 2006 by M. Gerdts

Page 135: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 135/243

5.4. LINEARIZED IMPLICIT RUNGE-KUTTA METHODS 127

Proof. Application of the LRK method to (5.4.2) and (5.4.8) yields (5.4.9) and

ky

ikz

i =

f (ν i) + hs

j=1

aij

f x(ν i)(ky

j − f (zm, ym)) + f t(ν i)(kz j − 1)

1

, i = 1, . . . , s ,

respectively, where

ν i =

zm + h

sl=1

ail · 1, ym + hs

l=1

ailf (zm, ym)

, i = 1, . . . , s .

kyi is equal to ki in (5.4.9), if ν i = ν i and zm = tm hold for all i and all m. The latter is fulfilled

if s

i=1 bi = 1 holds, since then zm+1 = zm + hs

i=1 bikzi = zm + h

si=1 bi = zm + h = tm+1.

The first condition requires that zm + hs

l=1 ail = tm + cih, which is satisfied if (5.4.10) holds.

In matrix notation, equation (5.4.9) is given by

(I − hB2(tm, xm, h)) k = c2(tm, xm, h), (5.4.11)

where

B2(tm, xm, h) :=

a1 ⊗ f x(tm + c1h, xm + hc1f (tm, xm))

...as ⊗ f x(tm + csh, xm + hcsf (tm, xm))

,

c2(tm, xm, h) := −hB2(tm, xm, h) · (e ⊗ f (tm, xm))

+f (tm + c1h, xm + hc1f (tm, xm))

.

..f (tm + csh, xm + hcsf (tm, xm))

,

(5.4.12)

and ai denotes the ith row of A = (aij) and e = (1, . . . , 1) ∈ Rs. Notice, that we exploited(5.4.10).

Lemma 5.4.3 allows to restrict the discussion to autonomous initial value problems (5.4.2), wheref does not depend explicitly on t. This simplifies the Taylor expansions considerably. From nowon, it is assumed, that the assumptions of Lemma 5.4.3 are satisfied. We determine the orderof consistency of the LRK method (5.4.9).The Taylor expansion of the exact solution of the autonomous problem is given by

x(tm + h) − x(tm)h

= f + h2!

f f + h2

3!f (f, f ) + f f f

+h3

4!

f (f , f , f ) + 3f (f f, f ) + f f (f, f ) + f f f f

+

h4

5!

f (4)(f , f , f , f ) + 6f (f f , f , f ) + 4f (f (f, f ), f )

+4f (f f f, f ) + 3f (f f, f f ) + f f (f , f , f )

+3f f (f f, f ) + f f f (f, f ) + f f f f f

+ O(h5),

where f is evaluated at x(tm). Notice, that we omitted brackets if no confusion was possible,e.g. we wrote f f f instead of f (f (f )).

c 2006 by M. Gerdts

Page 136: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 136/243

128 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

Now we derive the Taylor expansion of the approximate solution. The stages k i of the linearizedRunge-Kutta method (5.4.9) for the autonomous ODE (5.4.2) under consideration of (5.4.10)satisfy

ki = hs

j=1

aijf (xm + hcif )(k j ) + f (xm + hcif ) − hs

j=1

aijf (xm + hcif )(f )

= hs

j=1

aijf (xm + hcif )(k j ) + f (xm + hcif ) − hcif (xm + hcif )(f )

= h

pk=0

hkcki

k!

s j=1

aij f (k+1)n (k j , f , . . . , f

k−fold

)

+ p

k=0

hkcki

k! f (k)

n (f , . . . , f ) k−fold

− pk=0

hk+1ck+1i

k! f (k+1)

n (f , . . . , f ) (k+1)−fold

+O(h p+1)

=

pk=1

hkck−1i

(k − 1)!

s j=1

aijf (k)n (k j , f, . . . , f

(k−1)−fold

)

+

pk=0

hkcki

k! f (k)

n (f , . . . , f ) k−fold

− p

k=1

hkcki

(k − 1)!f (k)

n (f , . . . , f ) k−fold

+O(h p+1)

=

p

k=1

hkck−1i

(k − 1)!

s

j=1

aijf (k)n (k j , f, . . . , f

(k−1)−fold

)

+f +

pk=1

hkcki

1 − k

k! f (k)

n (f , . . . , f ) k−fold

+O(h p+1).

In particular we get for p = 4:

ki = f + hs

j=1

aijf (k j) + h2

ci

s j=1

aij f (k j , f ) − c2i

2 f (f, f )

+h3

c2i

2

s j=1

aijf (k j , f , f ) − c3i

3 f (f , f , f )

+h4

c3i6

s

j=1

aijf (4)(k j, f , f , f ) − c4i8

f (4)(f , f , f , f )

+ O(h5).

c 2006 by M. Gerdts

Page 137: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 137/243

5.4. LINEARIZED IMPLICIT RUNGE-KUTTA METHODS 129

Recursive evaluation leads to

ki = f + hs

j=1

aij f f + hs

l=1

a jl f (kl)

+h2

c j

sl=1

a jl f (kl, f ) − c2 j2

f (f, f )

+h3

c2 j

2

sl=1

a jl f (kl, f , f ) − c3 j3

f (f , f , f )

+ O(h4)

+h2

ci

s j=1

aijf

f + h

sl=1

a jlf (kl)

+h

2c j

s

l=1 a jl f

(kl, f ) −c2 j

2 f

(f, f ) + O(h

3

), f −c2i

2 f (f, f )

+h3

c2

i

2

s j=1

aijf

f + h

sl=1

a jl f (kl) + O(h2), f , f

− c3

i

3 f (f , f , f )

+h4

c3

i

6

s j=1

aijf (4)(f , f , f , f ) − c4i

8 f (4)(f , f , f , f )

+ O(h5).

Sorting for powers of h and exploiting c i = s

j=1 aij yields

ki = f + hcif (f ) + h2

s

j=1

aij

sl=1

a jl f (f (kl)) + c2

i

2 f (f, f )

+h3

s

j=1

aijc j

sl=1

a jl f (f (kl, f )) −s

j=1

aij

c2 j

2 f (f (f, f ))

+ci

s j=1

aij

sl=1

a jlf (f (kl), f ) + c3i

6 f (f , f , f )

+h4

s

j=1

aij

c2 j

2

sl=1

a jl f (f (kl, f , f )) −s

j=1

aij

c3 j

3 f (f (f , f , f ))

+ci

s j=1

aijc j

sl=1

a jlf (f (kl, f ), f ) − ci

s j=1

aij

c2 j

2 f (f (f, f ), f )

+c2

i

2

s

j=1

aij

s

l=1

a jl f (f (kl), f , f ) + c4

i

24f (4)(f , f , f , f )

+ O(h5).

c 2006 by M. Gerdts

Page 138: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 138/243

130 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

Again, a recursive evaluation leads to

ki = f + hcif (f ) + h2

s

j=1

aij

s

l=1

a jl f (f f + hclf (f )

+h2

s

m=1

alm

sr=1

amrf (f (kr)) + c2

l

2 f (f, f ) + O(h3)

+

c2i

2 f (f, f )

+h3

s

j=1

aijc j

sl=1

a jl f (f

f + hclf (f ) + O(h2), f

) −s

j=1

aij

c2 j2

f (f (f, f ))

+ci

s j=1

aij

sl=1

a jl f (f

f + hclf (f ) + O(h2)

, f ) + c3i

6 f (f , f , f )

+h4 s j=1

aij c

3

j6

f (f (f , f , f )) + ci

s j=1

aij c

2

j2

f (f (f, f ), f )

+c2

i

2

s j=1

aijc jf (f (f ), f , f ) + c4i24

f (4)(f , f , f , f )

+ O(h5).

Sorting for powers of h and exploiting c i = s

j=1 aij yields

ki = f + hcif (f ) + h2s

j=1

aijc jf (f (f )) + c2

i

2 f (f, f )

+h3

s

j=1

aij

sl=1

a jl clf (f (f (f ))) +s

j=1

aij

c2 j2

f (f (f, f ))

+ci

s j=1

aij c jf (f (f ), f ) + c3

i

6 f (f , f , f )

+h4

s

j=1

aij

c3 j6

f (f (f , f , f )) + ci

s j=1

aij

c2 j

2 f (f (f, f ), f )

+c2i2

s j=1

aij c j f (f (f ), f , f ) + c4i24

f (4)(f , f , f , f )

+s

j=1

aij

sl=1

a jl

sm=1

almcmf (f (f (f (f ))))

+s

j=1

aij

sl=1

a jlc2

l

2 f (f (f (f, f ))) +

s j=1

aij c j

sl=1

a jl clf (f (f (f ), f ))

+ci

s

j=1

aij

s

l=1

a jl clf (f (f (f )), f )

+ O(h5).

c 2006 by M. Gerdts

Page 139: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 139/243

5.4. LINEARIZED IMPLICIT RUNGE-KUTTA METHODS 131

Finally, we obtain the expansion

si=1

biki =s

i=1

bif + hs

i=1

bicif f + h2s

i=1

bi

s

j=1

aijc j f f f + c2

i

2 f (f, f )

+h3

si=1

bi

s j=1

aij

sl=1

a jl clf f f f +s

j=1

aij

c2 j2

f f (f, f )

+ci

s j=1

aijc j f (f f, f ) + c3i

6 f (f , f , f )

+h4s

i=1

bi

s

j=1

aij

c3 j6

f f (f , f , f ) + ci

s j=1

aij

c2 j2

f (f (f, f ), f )

+c2i

2

s

j=1

aij c jf (f f , f , f ) + c4

i

24

f (4)(f , f , f , f )

+s

j,l,m=1

aija jl almcmf f f f f +s

j,l=1

aij a jlc2l2

f f f (f, f )

+s

j,l=1

aijc ja jl clf f (f f, f ) + ci

s j,l=1

aija jl clf (f f f, f )

+ O(h5).

An investigation of the local discretization error

lN = x(tm + h) − x(tm)

h −

s

i=1

biki

reveals that the terms involving the derivative f (f f, f f ) are missing. Hence, the maximalattainable order of consistency is 4. We thus proved

Theorem 5.4.4 Let f possess continuous and bounded partial derivatives up to order four and let (5.4.10) be valid. The LRK method (5.4.9) is consistent of order 1, if

si=1

bi = 1

holds. It is consistent of order 2, if in addition

si=1

bici =

1

2

holds. It is consistent of order 3, if in addition

si,j=1

biaijc j = 1

6,

si=1

bic2i = 1

3

hold. It is consistent of order 4, if in addition s

i,j,l=1

biaija jl cl = 1

24,

si,j=1

biaijc2 j =

1

12,

si,j=1

biciaijc j = 1

8,

si=1

bic3i = 1

4.

hold. p = 4 is the maximal attainable order of consistency.

c 2006 by M. Gerdts

Page 140: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 140/243

132 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

Remark 5.4.5 The above conditions are the usual conditions known for general IRK methods.It has to be mentioned, that the LRK method is A-stable if the nonlinear IRK method (5.3.2)is A-stable, since A-stability is defined for linear differential equations. For linear differential equations, the LRK method coincides with the IRK method.

In view of the completion of the convergence proof, we finally show the Lipschitz continuityof the increment function Φ for the two methods under suitable differentiability assumptions.Notice, that the LRK methods defined by (5.4.5) and (5.4.9) are one-step methods and we mayapply Theorem 5.1.8. Notice, that implicit methods can be viewed as one-step methods as well,provided that the arising nonlinear resp. linear equations can be solved.

Theorem 5.4.6 Let f be twice continuously differentiable. Then, Φ in (5.3.3) with k from (5.4.6) respectively (5.4.11) is locally Lipschitz continuous in the exact solution x.

Proof. Since Φ is a linear combination of ki, i = 1, . . . , s, it suffices to show the local Lipschitzcontinuity of k. Since f is twice continuously differentiable, the functions B i and ci, i = 1, 2, in

(5.4.7) and (5.4.12) are continuously differentiable w.r.t. t, x and h. In addition, for i = 1, 2 itholds

limh→0

Bi(tm, xm, h) = A ⊗ f x(tm, xm),

limh→0

ci(tm, xm, h) = e ⊗ f (tm, xm).

Hence, for sufficiently small h it holds hBi(ηn) < 1, i = 1, 2. Neumann’s lemma yields thenon-singularity of I −hBi, i = 1, 2 for all sufficiently small h. In either case, the implicit functiontheorem yields the existence of a continuously differentiable function k = k(tm, xm, h) satisfying(5.4.6) respectively (5.4.11). Application of the mean-value theorem for vector functions yields

k(tm, x, h) − k(tm, x, h) = 10

kx(tm, x + t(x − x), h)(x − x)dt ≤ Lδx − x,

where Lδ := maxζ −x(t)≤δ kx(t ,ζ,h) exists for some δ > 0 and all t ∈ [t0, tf ] and all sufficientlysmall h.

Exploiting Theorems 5.1.7, 5.1.8 we proved

Theorem 5.4.7 Let the assumptions of Theorems 5.4.2, 5.4.6 and Theorems 5.4.4, 5.4.6, re-spectively, be valid. Then the linearized LRK methods (5.4.6) and (5.4.11), respectively, are convergent. The order of convergence is identical with the order of consistency of the respective method.

Summarizing, the order of convergence of the LRK method depends on the choice of the initialguess for one step of Newton’s method. An appropriate choice allows to create methods up toorder 4. The methods are easy to implement and show a very good computational performancewithin the numerical solution of dynamic optimization problems. The following numerical ex-amples show that the use of LRK methods reduces the computation time considerably comparedto standard BDF integration schemes, while the accuracy remains the same.

5.4.3 Numerical Experiments for DAEs

Unfortunately, the theory for DAEs is not yet completed. Nevertheless, numerical tests showvery promising results, cf. Gerdts [Ger04]. For implicit DAEs the problem of choosing the initial

guess k[0] = (k[0]

1

, . . . , k[0]s ) in (5.4.1) appropriately arises. Again, we consider the integration step

c 2006 by M. Gerdts

Page 141: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 141/243

Page 142: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 142/243

Page 143: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 143/243

Page 144: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 144/243

136 CHAPTER 5. DISCRETIZATION METHODS FOR ODES AND DAES

c 2006 by M. Gerdts

Page 145: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 145/243

Page 146: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 146/243

138 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

We consider

Problem 6.1 (DAE Optimal Control Problem)

Minimize ϕ(x(t0), x(tf ))

s.t. DAE (6.1),

ψ(x(t0), x(tf )) = 0nψ,

c(t, x(t), u(t)) ≤ 0nc a.e. in [t0, tf ],

s(t, x(t)) ≤ 0ns in [t0, tf ].

Some remarks are in order.

Remark 6.2 In several aspects the statement of the optimal control problem 6.1 and particularly the DAE (6.1) is too general. First of all, from Chapter 4 we may draw the conclusion that control and algebraic variables have the same properties (both are L∞-functions). Hence, it makes

no sense – at least theoretically – to allow that ϕ and ψ in Problem 6.1 depend on pointwise evaluations of the algebraic variables. Secondly, the numerical computation of consistent initial values, cf. Definition 1.7, only works efficiently for certain subclasses of the DAE (6.1), e.g.Hessenberg DAEs according to Definition 1.5. Thirdly, as it has been mentioned already in Chapter 5 numerical methods again only work well for certain subclasses of (6.1), e.g. Hessenberg DAEs up to index 3.

6.1 Direct Discretization Methods

Direct discretization methods are based on a discretization of the infinite dimensional optimalcontrol problem. The resulting discretized problem will be a finite dimensional optimizationproblem. All subsequently discussed methods work on the (not necessarily equidistant) grid

GN := t0 < t1 < .. . < tN = tf (6.1.1)

with step sizes h j = t j+1 − t j , j = 0, . . . , N − 1 and mesh-size h := max j=0,...,N −1

h j . Often, GN will

be an equidistant partition of the interval [t0, tf ] with constant step size h = (tf − t0)/N andgrid points ti = t0 + ih, i = 0, . . . , N .A direct discretization method is essentially defined by the following operations:

• Control discretization:The control space L∞([t0, tf ],R

nu) is replaced by some M -dimensional subspace

L∞M ([t0, tf ],R

nu) ⊂ L∞([t0, tf ],Rnu)

where M ∈ N is finite. The dimension M usually depends on the number N of intervalsin GN , i.e. M = O(N ).

Let B := B1, . . . , BM be a basis of L∞M ([t0, tf ],R

nu). Then, every uM ∈ L∞M ([t0, tf ],R

nu)is defined by

uM (·) =M

i=1

wiBi(·) (6.1.2)

with coefficients w := (w1, . . . , wM ) ∈ RM . The dependence on the vector w is indicated

by the notationuM (t) = uM (t; w) = uM (t; w1, . . . , wM ). (6.1.3)

Furthermore, we may identify uM and w.

c 2006 by M. Gerdts

Page 147: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 147/243

6.1. DIRECT DISCRETIZATION METHODS 139

• State discretization:The DAE is discretized by a suitable discretization scheme, e.g. a one-step method (5.1.3)or a multi-step method (5.2.3).

• Constraint discretization:The control/state constraints are only evaluated on the grid GN .

• Optimizer:The resulting discretized optimal control problem has to be solved numerically, e.g. by aSQP method, cf. Section 3.8.2.

• Calculation of gradients:Most optimizers need the gradient of the objective function and the Jacobian of the con-straints. There are basically three different methods for calculation: Approximation byfinite differences, sensitivity equation approach, and adjoint equation approach.

In the sequel we illustrate the direct discretization method for a generic one-step method (5.1.3)which is supposed to be appropriate for the DAE (6.1). For a given (consistent) initial value x 0

and a given control approximation uM as in (6.1.2) and (6.1.3) the one-step method generatesvalues

xN (t0) = x0,xN (t j+1) = xN (t j ) + h jΦ(t j, xN (t j), w , h j ), j = 0, 1, . . . , N − 1.

Notice, that Φ in general depends in a nonlinear way on the control parametrization w. Wedistinguish two approaches for the discretization of Problem 6.1: the full discretization approach and the reduced discretization approach . But before discussing them in detail, we have to addressthe problem of calculating a consistent initial value x0.

6.1.1 Projection Method for Consistent Initial ValuesSince we are dealing with DAEs only consistent initial values x0 are admissible, cf. Definition 1.7.There are basically two approaches to tackle the computation of consistent initial values, cf.Gerdts [Ger01a, Ger03a] and Gerdts and Buskens [GB02]. The first approach was suggestedby Schulz et al. [SBS98]. This approach relaxes the algebraic constraints of the DAE in sucha way that the previously inconsistent value becomes consistent for the relaxed problem. Forthis approach it is necessary to introduce additional equality constraints to the optimizationproblem.

The second approach to be discussed in more detail projects inconsistent values onto consistentones. In order to define a suitable projection we assume that the DAE (6.1) is a HessenbergDAE of order k as in (1.14) in Definition 1.5, i.e. F is given by

x1(t) = f 1(t, x0(t), x1(t), x2(t), . . . , xk−2(t), xk−1(t), u(t)),

x2(t) = f 2(t, x1(t), x2(t), . . . , xk−2(t), xk−1(t), u(t)),

... . . .

xk−1(t) = f k−1(t, xk−2(t), xk−1(t), u(t)),

0ny = g(t, xk−1(t), u(t)).

Recall, that the component y = x0 is the algebraic variable and xd = (x1, . . . , xk−1) ∈Rnd is the differential variable. According to Definition 1.7 a consistent initial value x =

c 2006 by M. Gerdts

Page 148: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 148/243

140 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

(x0, x1, . . . , xk−1) ∈ Rnx for the Hessenberg DAE has to satisfy not only the algebraic con-straint but also the hidden constraints, i.e. x for j = 0, 1, . . . , k − 2 has to satisfy

G j(xd, w) := g( j)

t0, xk−1− j, . . . , xk−1, uM (t0; w), uM (t0; w), . . . , u

( j)M (t0; w)

= 0ny (6.1.4)

and

Gk−1(y, xd, w) := g( j)

t0, y , xd, uM (t0; w), uM (t0; w), . . . , u( j)M (t0; w)

= 0ny (6.1.5)

Notice, that only Gk−1 depends on the algebraic variable y while G j, j = 0, . . . , k − 2 do notdepend on y. This suggests to split the computation of a consistent initial value into twoparts. In the first part only the differential component xd is computed consistently with theconstraints G j , j = 0, . . . , k − 2. Afterwards, the corresponding algebraic variable consistentwith the equation Gk−1 is computed.For a given control parametrization w and a given estimate xd

0 of a consistent differential com-ponent (later on this will be an iterate of a SQP method) the closest consistent xd is given by

the nonlinear least-squares projection

X d0 (xd0, w) := arg min

z∈Rndz − xd

02 | G j (z, w) = 0ny , j = 0, . . . , k − 2. (6.1.6)

This is a parametric nonlinear optimization problem with parameters xd0 and w as in Section 3.7.

Under the assumptions of Theorem 3.7.3 the derivatives (X d0 )x(xd0, w) and (X d0 )w(xd

0, w) resultfrom a sensitivity analysis of (6.1.6) w.r.t. xd

0 and w and are given by the linear equation (3.7.1).This completes the first part of the computation of a consistent initial value.Now we focus on the equation (6.1.5). Thus, having computed X d0 (xd

0, w) the correspondingconsistent component y is given implicitly by

Gk−1(X d0 (xd0, w), y , w) = 0ny .

Recall, that the matrix

∂yGk−1(y, xd, w) = g xk−1

(·) · f k−1,xk−2(·) · · · f 2,x1(·) · f 1,x0(·),

cf. (1.15), is supposed to be non-singular since the DAE is Hessenberg of order k. By the implicitfunction theorem there exists a function Y : Rnd × RM → Rny with

Gk−1(Y (xd, w), xd, w) = 0ny

for all (xd, w) in some neighborhood of (X d0 (xd0, w), w). Differentiation w.r.t. xd and w andexploitation of the nonsingularity of G

k−1,y yields

Y x(xd, w) = −

Gk−1,y(y, xd, w)

−1G

k−1,x(y, xd, w),

Y w(xd, w) = −

Gk−1,y(y, xd, w)

−1G

k−1,w(y, xd, w)

at y = Y (xd, w). Differentiation of Y 0(xd0, w) := Y (X d0 (xd

0, w), w) w.r.t. xd0 and w yields

Y 0,x(xd0, w) = Y x(X d0 (xd

0, w), w) · (X d0 )x(xd0, w),

Y

0,w

(xd

0

, w) = Y

x

(X d

0

(xd

0

, w), w)·

(X d

0

)

w

(xd

0

, w) + Y

w

(X d

0

(xd

0

, w), w).

c 2006 by M. Gerdts

Page 149: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 149/243

6.1. DIRECT DISCRETIZATION METHODS 141

Let us summarize the results. For Hessenberg DAEs a method for the computation of theconsistent initial value

x(t0) = X 0(xd0, w) :=

Y 0(xd

0, w)X d0 (xd

0, w)

was developed. The derivatives X 0,x, X

0,w, Y

0,x, Y

0,w result from a sensitivity analysis.

Concerning general DAEs (6.1) the projection method will work whenever it is possible toidentify the algebraic constraints and the hidden constraints. Attempts towards more generalDAEs can be found in Leimkuhler et al. [LPG91] and Buskens and Gerdts [BG05]. From nowon it is assumed that there exists a method to compute a consistent initial value X 0(x0, w) forthe general DAE (6.1) for given initial guess x0 and control parametrization w. The functionX 0 is assumed to be at least continuously differentiable.

6.1.2 Full Discretization Approach

We obtain a discretization of the optimal control problem by replacing the DAE by the one-stepmethod and discretizing the constraints on the grid GN :

Problem 6.1.1 (Full Discretization)

Find a grid function xN : GN → Rnx, ti → xN (ti) =: xi and w ∈ RM such that the objective function

ϕ(x0, xN )

is minimized subject to

X 0(x0, w) − x0 = 0nx ,x j + h jΦ(t j, x j , w , h j ) − x j+1 = 0nx , j = 0, 1, . . . , N − 1,

ψ(x0, xN ) = 0nψ,

c(t j , x j , uM (t j; w)) ≤ 0nc , j = 0, 1, . . . , N ,s(t j, x j) ≤ 0ns , j = 0, 1, . . . , N .

Problem 6.1.1 is a finite dimensional optimization problem of type

minz

F (z) s.t. G(z) ≤ Θ, H (z) = Θ (6.1.7)

with the nx(N + 1) + M dimensional optimization variable

z := (x0, x1, . . . , xN , w),

the objective functionF (z) := ϕ(x0, xN ),

the (nc + ns) · (N + 1) inequality constraints

G(z) :=

c(t0, x0, uM (t0; w))...

c(tN , xN , uM (tN ; w))s(t0, x0)

...s(t

N , x

N )

,

c 2006 by M. Gerdts

Page 150: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 150/243

142 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

and the nx · (N + 1) + nψ equality constraints

H (z) :=

X 0(x0, w) − x0

x0 + h0Φ(t0, x0, w , h0)

−x1

...xN −1 + hN −1Φ(tN −1, xN −1, w , hN −1) − xN

ψ(x0, xN )

.

Structure of the Optimization Problem:The size of Problem (6.1.7) depends on nx, N , and M and can become very large. In practicedimensions up to a million of optimization variables or even more are not unrealistic. In orderto handle these dimensions it is essential to exploit the sparse structure of (6.1.7), cf. Betts and

Huffman [BH92, BH99]. On the other hand, an advantage of the full discretization approach isthat it is easy to compute derivatives w.r.t. the optimization variables. It holds:

F (z) =

ϕx0

Θ · · · Θ ϕxf

Θ · · · Θ , (6.1.8)

G(z) =

cx[t0] cu[t0] · uM,w (t0; w)

. . . . . .

cx[tN ] cu[tN ] · uM,w (tN ; w)

sx[t0]

. . . Θ

sx[tN ]

,(6.1.9)

H (z) =

X 0,x0 − I nx X 0,w

M 0 −I nx h0Φw[t0]

. . . ...

M N −1 −I nx hN −1Φw[tN −1]

ψx0 ψ

xf Θ

(6.1.10)

where M j := I nx + h j Φx(t j , x j, w , h j ), j = 0, . . . , N − 1.

6.1.3 Reduced Discretization Approach

The reduced discretization approach is also based on the full discretization 6.1.1. But in thereduced approach the equations resulting from the one-step method are not explicitly taken asequality constraints in the discretized optimization problem. It is clear that the variable x i+1 is

c 2006 by M. Gerdts

Page 151: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 151/243

6.1. DIRECT DISCRETIZATION METHODS 143

completely defined by (ti, xi, w). Solving the equations recursively we find:

x0 = X 0(x0, w),x1 = x0 + h0Φ(t0, x0, w , h0)

= X 0(x0, w) + h0Φ(t0, X 0(x0, w), w , h0)

=: X 1(x0, w),x2 = x1 + h1Φ(t1, x1, w , h1)

= X 1(x0, w) + h1Φ(t1, X 1(x0, w), w , h1)=: X 2(x0, w),...

xN = xN −1 + hN −1Φ(tN −1, xN −1, w , hN −1)= X N −1(x0, w) + hN −1Φ(tN −1, X N −1(x0, w), w , hN −1)=: X N −1(x0, w).

(6.1.11)

Of course, this is just the formal procedure of solving the DAE for given consistent initial valueand control parametrization w. We obtain:

Problem 6.1.2 (Reduced Discretization)Find x0 ∈ Rnx and w ∈ RM such that the objective function

ϕ(X 0(x0, w), X N (x0, w))

is minimized subject to

ψ(X 0(x0, w), X N (x0, w)) = 0nψ,

c(t j, X j(x0, w), uM (t j; w)) ≤ 0nc, j = 0, 1, . . . , N ,s(t j, X j (x0, w))

≤ 0ns, j = 0, 1, . . . , N .

Again, Problem 6.1.2 is a finite dimensional nonlinear optimization problem of type (6.1.7) withthe nx + M optimization variables z := (x0, w), the objective function

F (z) := ϕ(X 0(x0, w), X N (x0, w)), (6.1.12)

the (nc + ns) · (N + 1) inequality constraints

G(z) :=

c(t0, X 0(x0, w), uM (t0; w))...

c(tN , X N (x0, w), uM (tN ; w))

s(t0, X 0(x0, w))...

s(tN , X N (x0, w))

, (6.1.13)

and the nψ equality constraints

H (z) := ψ(X 0(x0, w), X N (x0, w)). (6.1.14)

Structure of the Optimization Problem:The size of Problem 6.1.2 is small compared to Problem 6.1.1 since we eliminated most of the

c 2006 by M. Gerdts

Page 152: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 152/243

144 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

equality constraints as well as the variables x2, . . . , xN . Only the initial value x0 remains asan optimization variable. Unfortunately the derivatives are not sparse anymore and it is moreinvolved to compute them:

G(z) =

cx[t0]

·X 0,x0

cx[t0]

·X 0,w + cu[t0]

·uM,w (t0; w)

cx[t1] · X 1,x0 cx[t1] · X 1,w + cu[t1] · uM,w (t1; w)

... ...

cx[tN ] · X N,x0 cx[tN ] · X N,w + cu[tN ] · uM,w(tN ; w)

sx[t0] · X 0,x0 sx[t0] · X 0,w

sx[t1] · X 1,x0 sx[t1] · X 1,w

... ...

sx[tN ]·

X

N,x0

sx[tN ]·

X

N,w

, (6.1.15)

H (z) =

ψx0 · X 0,x0 + ψ

xf · X N,x0

ψx0 · X 0,w + ψ

xf · X N,w

. (6.1.16)

Herein, the sensitivities

X i,x0(x0, w), X i,w(x0, w), i = 0, . . . , N (6.1.17)

have to be computed. This problem will be addressed later on in detail.

Remark 6.1.3 Based on the same ideas it is possible and often necessary to extend the reduced discretization method to a reduced multiple shooting method by introducing multiple shooting

nodes. Details can be found in Gerdts [Ger01a, Ger03a]. Related methods based on the multiple shooting idea are discussed in, e.g., Deuflhard [Deu74], Bock [BP84], Stoer and Bulirsch [SB90],Leineweber [Lei95], Schulz et al. [SBS98], Hinsberger [Hin97], Oberle and Grimm [OG01].

6.1.4 Control Discretization

We address the problem of discretizing the control and specify the finite dimensional sub-space L∞

M ([t0, tf ],Rnu) ⊂ L∞([t0, tf ],R

nu). We intend to replace the control variable u(·) ∈L∞([t0, tf ],R

nu) by a B-spline representation. Given k ∈ N the elementary B-splines B ki (·) of

order k , i = 1, . . . , N + k − 1, are defined by the recursion

B1i (t) :=

1, if τ i ≤ t < τ i+1

0, otherwise ,

Bki (t) :=

t − τ iτ i+k−1 − τ i

Bk−1i (t) +

τ i+k − t

τ i+k − τ i+1Bk−1

i+1 (t)

(6.1.18)

on the auxiliary grid

GkN := τ i | i = 1, . . . , N + 2k − 1 (6.1.19)

with auxiliary grid points

τ i :=

t0, if 1 ≤ i ≤ k,ti−k, if k + 1 ≤ i ≤ N + k − 1,t

N , if N + k

≤ i ≤

N + 2k−

1.

c 2006 by M. Gerdts

Page 153: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 153/243

6.1. DIRECT DISCRETIZATION METHODS 145

Notice, that the convention 0/0 = 0 is used if auxiliary grid points coincide in the recur-sion (6.1.18). The evaluation of the recursion (6.1.18) is well-conditioned, cf. Deuflhard andHohmann [DH91]. The elementary B-splines B k

i (·), i = 1, . . . , N +k−1 restricted to the intervals[t j, t j+1], j = 0, . . . , N − 1 are polynomials of degree at most k − 1, i.e.

Bki (·)

[tj ,tj+1] ∈ P k−1([t j , t j+1]),

and they form a basis of the space of splines

s(·) ∈ C k−2([t0, tf ],R)

s(·)[tj ,tj+1]

∈ P k−1([t j , t j+1]) for j = 0, . . . , N − 1

.

Obviously, the elementary B-splines are essentially bounded, i.e. B ki (·) ∈ L∞([t0, tf ],R).

For k

≥ 2 it holds B k

i (

·)

∈ C k−2([t0, tf ],R) and for k

≥ 3 the derivative obeys the recursion

d

dtBk

i (t) = k − 1

τ i+k−1 − τ iBk−1

i (t) − k − 1

τ i+k − τ i+1Bk−1

i+1 (t).

Furthermore, the elementary B-splines possess local support supp(B ki ) ⊂ [τ i, τ i+k]. More pre-

cisely, for k > 1 it holds

B1i (t)

> 0, if t ∈ (τ i, τ i+k),= 0, otherwise.

The cases k = 1 or k = 2 frequently occur in numerical computations. For k = 1 the elemen-tary B-splines are piecewise constant functions, while for k = 2 we obtain the continuous andpiecewise linear functions

B2i (t) =

t − τ iτ i+1 − τ i

, if τ i ≤ t < τ i+1,

τ i+2 − t

τ i+2 − τ i+1, if τ i+1 ≤ t < τ i+2,

0, otherwise.

Similarly, in some situations it might be necessary to have a continuously differentiable function

as, e.g.,

B3i (t) =

(t − τ i)2

(τ i+2 − τ i)(τ i+1 − τ i), if t ∈ [τ i, τ i+1),

(t − τ i)(τ i+2 − t)

(τ i+2 − τ i)(τ i+2 − τ i+1) +

(τ i+3 − t)(t − τ i+1)

(τ i+3 − τ i+1)(τ i+2 − τ i+1), if t ∈ [τ i+1, τ i+2),

(τ i+3 − t)2

(τ i+3 − τ i+1)(τ i+3 − τ i+2), if t ∈ [τ i+2, τ i+3),

0, otherwise.

Figure 6.1 visualizes B-splines of orders k = 2, 3, 4.

c 2006 by M. Gerdts

Page 154: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 154/243

146 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

0

0.2

0.4

0.6

0.8

1

0.2 0.4 0.6 0.8 1

0

0.2

0.4

0.6

0.8

1

0.2 0.4 0.6 0.8 1

0

0.2

0.4

0.6

0.8

1

0.2 0.4 0.6 0.8 1

Figure 6.1: B-splines of order k = 2 (left), k = 3 (middle), and k = 4 (right)with [t0, tf ] = [0, 1] and N = 5 on an equidistant grid.

In view of the optimal control problem, we consider discretized controls

uM (·) ∈ L∞M ([t0, tf ],R

nu) :=

N +k−1

i=1

wi Bki (·)

wi ∈ Rnu , i = 1, . . . , N + k − 1

. (6.1.20)

Notice, that each uM (·) ∈ L∞M ([t0, tf ],R

nu) is determined by the M := nu(N + k − 1) con-trol parameters w := (w1, . . . , wN +k+1) ∈ Rnu(N +k−1) and hence L∞

M ([t0, tf ],Rnu) is a finite

dimensional subspace of L∞([t0, tf ],Rnu). The dependence on the vector w is indicated by the

notation

uM

(t) = uM

(t; w) = uM

(t; w1

, . . . , wN +k−1

).

The coefficients wi are known as de Boor points . In the special cases k = 1 and k = 2 the de Boorpoints wi can be interpreted as function values wi = uN (ti). As already mentioned, most oftenthe choices k = 1 (piecewise constant approximation) or k = 2 (continuous and piecewise linearapproximation) are preferred by many authors, cf., e.g., von Stryk [vS94], Buskens [Bus98], andBuskens and Gerdts [BG00].

The choice of B-splines has two major advantages from the numerical point of view. First, it iseasy to create approximations with prescribed smoothness properties. Second, the de Boor pointwi influences the function value uM (t) only in the interval t ∈ [τ i, τ i+k] due to the local support of Bk

i . This porperty will lead to certain sparsity patterns in the gradient of the objective functionand the Jacobian of the constraints. The exploitation of this sparsity reduces the computational

effort for the numerical solution considerably, cf. Gerdts [Ger01a, Ger03a].If instead an interpolating cubic spline is used as control approximation as in Kraft [KE85] theparametrization of the approximate control influences the function values in the whole timeinterval [t0, tf ] and the local influence of wi is lost. Besides the missing sparsity this may leadto numerical instabilities as well. Barclay et al. [BGR97] employ piecewise defined Hermitepolynomials as approximations.

Remark 6.1.4 Of course, it is possible to use individual B-splines of different order for each component of the control vector uM . For the sake of simplicity, we will not discuss these control approximations and restrict the discussion to the case where B-splines of the same order are used for every component of the control vector.

c 2006 by M. Gerdts

Page 155: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 155/243

Page 156: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 156/243

148 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

There is no convergence at all. The numerical solution for N = 40 shows a control u oscillating strongly between 0 at ti + h/2 and approximately −1/(2h) at ti:

-90

-80

-70

-60

-50

-40

-30

-20

-10

0

10

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Now we use a piecewise constant control approximation in the modified Euler’s method and obtain the following error in x in the norm · ∞:

N Error in x Order

10 0.3358800781952942E-03 –

20 0.8930396513584515E-04 1.911150140 0.2273822819465199E-04 1.973604480 0.5366500129055929E-05 2.0830664

160 0.1250729642299220E-05 2.1012115 320 0.7884779272826492E-06 0.6656277

For this control approximation the modified Euler’s method converges of second order!

6.2 Calculation of Gradients

The applicability of the SQP method for solving the discretized optimal control problems 6.1.1resp. 6.1.2 is straightforward. Nevertheless, the SQP method needs the gradient F of the ob-

jective function and the Jacobians G and H of the constraints. In case of the full discretization

approach these quantities are easily computed according to (6.1.8)-(6.1.10). In case of the re-duced discretization approach it is more involved to compute them. We will see that this canbe achieved in different ways:

• The sensitivity differential equation approach is advantageous if the number of constraintsis (much) larger than the number of variables in the discretized problem.

• The adjoint equation approach is preferable if the number of constraints is less than thenumber of variables in the discretized problem.

• A very powerful and user-friendly tool for the evaluation of gradients is algorithmic dif- ferentiation . This approach assumes that the evaluation of a function is performed by

c 2006 by M. Gerdts

Page 157: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 157/243

6.2. CALCULATION OF GRADIENTS 149

a Fortran or C procedure (or any other supported programming language). Algorith-mic differentiation means that the complete procedure is differentiated step by step usingroughly speaking chain and product rules. The result is again a Fortran or C proce-dure that provides the gradient of the function. Essentially the so-called ‘forward mode’in algorithmic differentiation corresponds to the sensitivity equation approach while the‘backward mode’ corresponds to the adjoint approach. Details can be found on the webpage www.autodiff.org and in the book of Griewank [Gri00] and in the review article of Griewank [Gri03].

• The approximation by finite differences is straightforward but has the drawback of beingcomputationally expensive and often suffers from low accuracy. Nevertheless, this approachis often used if some solver depending on optimization variables is used as a black box insidean algorithm.

We only concentrate on the first two approaches for calculating derivatives for the reduceddiscretization approach.

6.2.1 Sensitivity Equation Approach

We restrict the discussion to the one-step method

X i+1(z) = X i(z) + hiΦ(ti, X i(z), w , hi), i = 0, 1, . . . , N − 1. (6.2.1)

Recall that z = (x0, w) denotes the variables in the optimization problem 6.1.2. We intend tocompute the sensitivities

S i := ∂X i(z)

∂z , i = 0, 1, . . . , N .

For i = 0 it holds

S 0 = ∂X 0(z)∂z

∈ Rnx×(nx+M ), (6.2.2)

where X 0(z) is a continuously differentiable function that provides a consistent initial value.Differentiaton of (6.2.1) w.r.t. z yields the relationship

S i+1 = S i + hi

Φ

x(ti, X i(z), w , hi) · S i + Φw(ti, X i(z), w , hi) · ∂ w

∂z

(6.2.3)

for i = 0, 1, . . . , N − 1. This approach is known as internal numerical differentiation (IND), cf.Bock [Boc87]. The IND-approach is based on the differentiation of the discretization scheme(e.g. the one-step method) w.r.t. z. Computing the derivatives for the reduced problem 6.1.2

essentially amounts to solving one initial value problem of size nx · (1 + nx + M ). It is worthpointing out that the size depends on the number of unknowns in the programming problem,but it does not depend on the number of constraints.

In practice, the computation of the derivatives Φx and Φ

w is non-trivial. We illustrate thedifficulties for the implicit Euler’s method and consider the integration step t i → ti+1 = ti + hi.Discretization of (6.1) with X i ≈ x(ti), X i+1 ≈ x(ti+1) leads to the nonlinear equation

F

ti+1, X i+1,

X i+1 − X ihi

, uM (ti+1; w)

= 0nx . (6.2.4)

Provided that this nonlinear equation possesses a solution and that we may apply the implicitfunction theorem, this solution will depend on z, i.e. X i+1 = X i+1(z). Now, we need the

c 2006 by M. Gerdts

Page 158: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 158/243

150 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

derivative of the numerical solution X i+1(z) w.r.t. to z, i.e. S i+1. According to the implicitfunction theorem, differentiation of (6.2.4) w.r.t. z yields the linear equation

F x +

1

hiF x X i+1

· S i+1 = −F uX i+1

· uM,z (ti+1; w) + 1

hiF xX i+1

· S i. (6.2.5)

This formula can be obtained in a different way. Let x(t; z) denote the solution of the DAE (6.1)for given z. If F is sufficiently smooth and x is continuously differentiable w.r.t. t and z and itholds ∂

∂zdx(t;z)

dt = ddt

∂x(t;z)∂z , differentiation of (6.1) w.r.t. z results in a linear matrix DAE – the

sensitivity DAE F x[t] · S (t) + F x[t] · S (t) + F u[t] · uM,z (t; w) = 0nx×nx

for the sensitivity matrix S (t) := ∂x(t; z)/∂z. Discretization of the sensitivity DAE with thesame method as in (6.2.4) and for the same time step leads again to the linear equation (6.2.5).Hence, both approaches coincide, provided that X i+1 solves (6.2.4) exactly. In practice, this isnot the case and (6.2.4) is solved numerically by Newton’s method

F x + 1hi

F x

X (k)i+1

· ∆X (k)i+1 = −F

X (k)i+1

, X (k+1)i+1 = X (k)

i+1 + ∆X (k)i+1, k = 0, 1, 2, . . . .

Notice, that the iteration matrix F x + 1hi

F x at the last iterate can be re-used for computing S i+1.In practice, the iteration matrix within Newton’s method is kept constant for several Newtonsteps to speed up the procedure. But it is important to mention that the iteration matrix hasto be re-evaluated prior to the calculation of the sensitivity matrix S i+1. The common strategyof keeping the iteration matrix constant even for some integration steps leads to poor resultsconcerning the accuracy of the sensitivity matrix. Of course, if the iteration matrix is keptconstant or if only a fixed number of Newton steps are performed then the resulting S i+1 is onlyan approximation of the correct derivative ∂ X i+1(z)/∂z.

Remark 6.2.1

• The IND approach is applicable for multistep method such as BDF as well.

• Similar strategies are discussed in Caracotsios and Stewart [CS85], Maly and Petzold [MP96],Brenan et al. [BCP96], Heim [Hei92], and Kiehl [Kie98]. A comparison of different strate-gies can be found in Feehery et al. [FTB97].

6.2.2 Adjoint Equation Approach: The Discrete CaseThe adjoint method avoids the calculation of the sensitivities S i. We demonstrate the methodfor a prototype function of type

Γ(z) := γ (X 0(z), X N (z), z).

Obviously, ϕ, ψ and essentially c and s are of this type.We intend to derive a procedure for calculating Γ (z) subject to the difference equations

X i+1(z) − X i(z) − hiΦ(ti, X i(z), w , hi) = 0nx , i = 0, . . . , N − 1. (6.2.6)

The initial state X 0(z) ∈ Rnx is assumed to be sufficiently smooth as a function of z andconsistent with the underlying DAE for all z. We consider the auxiliary functional

J (z) := Γ(z) +N −1

i=0

λi+1 (X i+1(z) − X i(z) − hiΦ(ti, X i(z), w , hi))

c 2006 by M. Gerdts

Page 159: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 159/243

6.2. CALCULATION OF GRADIENTS 151

with multipliers λi, i = 1, . . . , N . Differentiating J w.r.t. z yields the expression

J (z) = γ x0 · S 0 + γ xN · S N + γ z

+N −1

i=0

λ

i+1S i+1

−S i

−hiΦ

x

[ti]·

S i−

hiΦ

w

[ti]· ∂ w

∂z = γ x0 · S 0 + γ xN

· S N + γ z

+N

i=1

λi S i −N −1i=0

λi+1

S i + hiΦ

x[ti] · S i + hiΦw[ti] · ∂ w

∂z

=

γ x0 − λ1 − h0λ1 Φx[t0]

· S 0 +

γ xN

+ λN

· S N + γ z

+N −1i=1

λi − λi+1 − hiλi+1Φ

x[ti]

· S i −N −1i=0

hiλi+1Φw[ti] · ∂ w

∂z .

The terms S i = ∂X i(z)/∂z are just the sensitivities in the sensitivity equation approach whichwe do not (!) want to compute here. Hence, we have to ensure that the expressions involvingS i are eliminated. This leads to the adjoint equation

λi − λi+1 − hiλi+1Φx[ti] = 0nx , i = 0, . . . , N − 1 (6.2.7)

and the transversality condition

λN = −γ xN (X 0(z), X N (z), z). (6.2.8)

Notice, that the adjoint equation is solved backwards in time. With these expressions thederivative of J reduces to

J (z) =

γ x0 − λ0

· S 0 + γ z −

N −1i=0

hiλi+1Φw[ti] · ∂w

∂z . (6.2.9)

Herein, the sensitivity matrix S 0 is given by

S 0 = X 0(z). (6.2.10)

It remains to show that J (z) = Γ(z) holds:

Theorem 6.2.2 It holds

Γ(z) = J (z) =

γ x0 − λ0

· S 0 + γ z −

N −1i=0

hiλi+1Φu[ti] · ∂ w

∂z . (6.2.11)

Proof. By multiplication of the sensitivity equation

S i+1 − S i − hiΦx[ti]S i − hiΦ

w[ti] · ∂ w

∂z = 0nx , i = 0, . . . , N − 1

with λi+1 from the left we obtain

−hiλi+1Φ

w[ti]

· ∂ w

∂z

=

−λi+1S i+1 + λi+1S i + hiλi+1Φ

x[ti]

·S i, i = 0, . . . , N

−1

c 2006 by M. Gerdts

Page 160: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 160/243

152 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

and hence

J (z) =

γ x0 − λ0

· S 0 + γ z

+

N −1

i=0 λ

i+1S i + hiλ

i+1Φ

x[ti]S i − λ

i+1S i+1

(6.2.7)=

γ x0 − λ0

· S 0 + γ z +

N −1i=0

λi S i − λi+1S i+1

=

γ x0 − λ0

· S 0 + γ z + λ0 S 0 − λN S N

(6.2.8)= γ x0S 0 + γ z + γ xN

S N

= Γ(z).

With Γ

(z) = J

(z) we finally found a formula for the gradient of Γ. Γ itself is a placeholder forthe functions F, G = (G1, . . . , Gm), H = (H 1, . . . , H p) in (6.1.12)-(6.1.14).In order to compute F , G, H of (6.1.12)-(6.1.14) for each (!) component of F,G, H an adjointequation with appropriate transversality condition has to be solved. This essentially correspondsto solving an initial value problem of dimension nx · (2 + m + p). In addition, the trajectory(X i, i = 0, . . . , N ) has to be stored. It is important to mention that the effort does not dependon the number M of control parameters! The method is particularly efficient if none or very fewconstraints are present.

6.2.2.1 Application to Runge-Kutta Methods

The function Φ in the preceding section is specified for Runge-Kutta methods and DAEs of type

F (t, x(t), x(t), uM (t; w)) = 0nx , x(t0) = X 0.

An implicit Runge-Kutta method with s stages is given by the iteration

X i+1 = X i + hΦ(ti, X i, w , h), i = 0, 1, . . . , N − 1,

where the increment function Φ is defined by

Φ(t,x,w,h) :=s

j=1

b j k j(t,x,w,h) (6.2.12)

and the stage derivatives k = (k1, . . . , ks) are implicitly defined by the nonlinear equation

G(k,t,x,w,h) :=

F

t + c1h, x + h

s j=1

a1 j k j , k1, uM (t + c1h; w)

F

t + c2h, x + h

s j=1

a2 j k j, k2, uM (t + c2h; w)

...

F

t + csh, x + hs

j=1

asjk j, ks, uM (t + csh; w)

= 0s·nx . (6.2.13)

c 2006 by M. Gerdts

Page 161: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 161/243

6.2. CALCULATION OF GRADIENTS 153

The nonlinear equation (6.2.13) is solved numerically by Newton’s method:

Gk(k( j), t ,x,w,h)∆k( j) = −G(k( j), t ,x,w,h),

k( j+1) = k( j) + ∆k( j), j = 0, 1, 2, . . . .

Under the assumption that the derivative G

k is non-singular at a solution k, the implicit functiontheorem yields the existence of the function k = k(t,x,w,h) satisfying

G(k(t,x,w,h), t ,x,w,h) = 0s·nx .

By differentiation w.r.t. x and w we find

kx(t,x,w,h) = − G

k(k,t,x,w,h)−1

Gx(k,t,x,w,h),

kw(t,x,w,h) = − G

k(k,t,x,w,h)−1

Gw(k,t,x,w,h),

where

Gk(k,t,x,w,h) =

ha11M 1 + T 1 ha12M 1 · · · ha1sM 1

ha21M 2 ha22M 2 + T 2 . . . .

..

... . . .

. . . has−1,sM s−1

has1M s · · · has,s−1M s hassM s + T s

,

M j := F x

t + c jh, x + h

sl=1

a jlkl, k j, uM (t + c jh; w)

, j = 1, . . . , s ,

T j := F x

t + c jh, x + h

sl=1

a jlkl, k j, uM (t + c jh; w)

, j = 1, . . . , s ,

Gx(k,t,x,w,h) =

F xt + c1h, x + h

s j=1

a1 j k j, k1, uM (t + c1h; w)

...

F x

t + csh, x + h

s j=1

asjk j, ks, uM (t + csh; w)

,

Gw(k,t,x,w,h) =

F u

t + c1h, x + h

s j=1

a1 jk j, k1, uM (t + c1h; w)

· uM,w(t + c1; w)

...

F u

t + csh, x + hs

j=1

asjk j , ks, uM (t + csh; w)

· uM,w (t + cs; w)

.

Notice, that (6.2.7) and (6.2.9) require the partial derivatives Φx(t,x,w,h) and Φ

w(t,x,w,h).According to (6.2.13) these values are given by

Φx(t,x,w,h) =

s j=1

b jk j,x(t,x,w,h),

Φw(t,x,w,h) =

s

j=1

b jk j,w(t,x,w,h).

c 2006 by M. Gerdts

Page 162: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 162/243

154 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

In particular, for the implicit Euler’s method we have

G(k,t,x,w,h) = F (t + h, x + hk,k,uM (t + h; w))

and

Φx(t,x,w,h) = kx(t,x,w,h) = −

hF x[t + h] + F x[t + h]−1 · F x[t + h],

Φw(t,x,w,h) = kw(t,x,w,h) = −

hF x[t + h] + F x[t + h]−1 · F u[t + h] · uM,w(t + h; w).

Example 6.2.3 (Explicit Runge-Kutta Methods)We briefly discuss the important subclass of ODEs

F (t, x(t), x(t), uM (t; w)) := x(t) − f (t, x(t), uM (t; w))

in combination with an s-staged explicit Runge-Kutta method

X i+1 = X i + h

s j=1

b jk j (ti, X i, w , h), i = 0, 1, . . . , N − 1,

where the stage derivatives k j = k j (t,x,w,h) are recursively defined by the equation

k j (t,x,w,h) := f

t + c j h, x + h

j−1=1

a j k, uM (t + c j h; w)

, j = 1, . . . , s . (6.2.14)

By differentiation w.r.t. x and w we find

k

j,x = f

xt + c j h, x + h

j−1

=1

a jk, uM (t + c j h; w)I + h

j−1

=1

a jk

,x ,

k j,w = f x

t + c j h, x + h

j−1=1

a jk, uM (t + c j h; w)

h

j−1=1

a jk,w

+f u

t + c jh, x + h

j−1=1

a jk, uM (t + c jh; w)

· uM,w(t + c j; w)

for j = 1, . . . , s and

Φx(t,x,w,h) =

s

j=1

b jk j,x(t,x,w,h),

Φw(t,x,w,h) =

s j=1

b jk j,w(t,x,w,h).

6.2.2.2 Simplecticity

The overall integration scheme consisting of the adjoint equation and the one-step methodhas a nice additional property – simplecticity, which was also observed by Laurent-Varin etal. [LVBB+04] for ODE optimal control problems. The overall scheme reads as

λi+1 = λi

−hΦ

x

(ti, X i, w , h)λi+1, X i+1 = X i + hΦ(ti, X i, w , h)

c 2006 by M. Gerdts

Page 163: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 163/243

6.2. CALCULATION OF GRADIENTS 155

Using the auxiliary function H (t,x,λ,w,h) := λΦ(t,x,w,h) the integration scheme can berewritten in the form

λi+1 = λi − hH x(ti, X i, λi+1, w , h), X i+1 = X i + hH λ(ti, X i, λi+1, w , h).

Solving this equation leads to λi+1

X i+1

=

Ψ1(λi, X i)Ψ2(λi, X i)

=

I + hΦ

x(ti, X i, w , h)−1

λi

X i + hΦ(ti, X i, w , h)

.

Hence, we obtain the integration scheme

Ψ1(λi, X i) + hH x(ti, X i, Ψ1(λi, X i), w , h) = λi,

Ψ2(λi, X i) − hH λ(ti, X i, Ψ1(λi, X i), w , h) = X i.

Differentiation w.r.t. (λi, X i) leads to I + h(H xλ) 0

−hH λλ I

Ψ

1,λiΨ

1,xi

Ψ2,λi

Ψ2,xi

=

I −hH xx

0 I + h(H λx)

Exploiting H λλ = 0 yields the Jacobian of Ψ = (Ψ1, Ψ2) to be

Ψ =

Ψ

1,λiΨ

1,xi

Ψ2,λi

Ψ2,xi

=

I + h(H xλ)

−1 −h

I + h(H xλ)−1

H xx

0 I + hH xλ

.

It is straightforward to show (Ψ )J Ψ = J , where

J = 0 I −I 0

.

Thus, we proved

Theorem 6.2.4 Ψ is symplectic, i.e. it holds (Ψ)J Ψ = J .

6.2.3 Adjoint Equation Approach : The Continuous Case

If the DAE is solved by an state-of-the-art integrator with automatic step-size and order selectionthen the adjoint equation approach as described before becomes very costly in view of the storageneeded by the method. Actually, the approximation at every time point generated by the step-size selection algorithm has to be stored. This is a potentially large number. In this case

it is more convenient not to consider state approximations obtained by the one-step method.Instead, the continuous DAE (6.1) for the discretized control uM (t; w) is viewed as a ‘continuous’constraint. This will lead to a continuous state x(t; z) depending on z. More precisely, the statex is given by the initial value problem

F (t, x(t; z), x(t; z), uM (t; w)) = 0nx , x(t0; z) = X 0(z). (6.2.15)

The initial state X 0(z) ∈ Rnx is assumed to be sufficiently smooth as a function of z andconsistent with the DAE for all z .We discuss the adjoint approach for calculating the gradient of the function

Γ(z) := γ (x(t0; z), x(tf ; z), z) (6.2.16)

c 2006 by M. Gerdts

Page 164: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 164/243

156 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

w.r.t. the parameter vector z ∈ Rnz . In order to derive an expression for the derivative Γ (z)the constraint (6.2.15) is coupled by use of the multiplier function λ to the function in (6.2.16)as follows

J (z) := Γ(z) + tf

t0

λ(t)F (t, x(t; z), x(t; z), uM (t; w)dt.

Differentiating J w.r.t. z yields the expression

J (z) = Γ(z) +

tf

t0

λ(t)

F x[t] · S (t) + F x[t] · S (t) + F u[t] · uM,z (t; w)

dt

= γ x0 · S (t0) + γ xf · S (tf ) + γ z

+

tf

t0

λ(t) · F x[t] · S (t) + λ(t) · F x[t] · S (t) + λ(t) · F u[t] · uM,z (t; w)dt.

Integration by parts yields

J (z) = γ xf + λ(tf )

·F x[tf ] ·

S (tf ) + γ x0 −

λ(t0)

·F x[t0] ·

S (t0) + γ z

+

tf

t0

λ(t) · F x[t] − d

dt

λ(t) · F x[t]

· S (t) + λ(t) · F u[t] · uM,z (t; w)dt.

λ is chosen such that the adjoint DAE

λ(t) · F x[t] − d

dt

λ(t) · F x[t]

= 0nx (6.2.17)

is satisfied, provided that such a function λ exists. Then, the derivative of J reduces to

J (z) =

γ xf + λ(tf )

· F x[tf ]

· S (tf ) +

γ x0 − λ(t0) · F x[t0]

· S (t0) + γ z

+ tf

t0λ(t) · F u[t] · uM,z (t; w)dt.

A connection with the sensitivity DAE arises as follows. Almost everywhere it holds

d

dt

λ(t) · F x[t] · S (t)

=

d

dt

λ(t) · F x[t]

· S (t) + λ(t) · F x(t) · S (t)

= λ(t)F x[t] · S (t) + λ(t)(−F x[t] · S (t) − F u[t] · uM,z (t; w))

= −λ(t)F u[t] · uM,z (t; w).

Using the fundamental theorem of calculus we obtain

λ(t) · F x[t] · S (t)

tf

t0= −

tf t0

λ(t)F u[t] · uM,z (t; w)dt. (6.2.18)

The gradient then becomes

J (z) =

γ xf + λ(tf )

· F x[tf ]

· S (tf ) +

γ x0 − λ(t0) · F x[t0]

· S (t0) + γ z

+

tf

t0

λ(t) · F u[t] · uM,z (t; w)dt

(6.2.18)= γ xf

S (tf ) + γ x0 · S (t0) + γ z

= Γ(z).

c 2006 by M. Gerdts

Page 165: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 165/243

6.2. CALCULATION OF GRADIENTS 157

Hence, J (z) is identical to Γ(z) provided that λ satisfies (6.2.17).While it is cheap to compute the sensitivity matrix

S (t0) = X 0(z), (6.2.19)

the sensitivity S (tf ) in the expression for J

(z) has to be eliminated by a proper choice of λ(tf )as we intend to avoid the explicit computation of S (tf ). Moreover, λ(tf ) has to be chosenconsistently with the adjoint DAE (6.2.17). We discuss the procedure of defining appropriateconditions for λ(tf ) for some common cases, cf. Cao et al. [CLPS03]:

(i) Let F x be non-singular a.e. in [t0, tf ]. Then, the DAE has index zero and λ(tf ) is obtainedby solving the linear equation

γ xf + λ(tf )

· F x[tf ] = 0nx (6.2.20)

and we findλ(tf )

= −γ xf F x[tf ]

−1

and thus

J (z) =

γ x0 − λ(t0) · F x[t0]

· S (t0) + γ z +

tf

t0

λ(t) · F u[t] · uM,z (t; w)dt. (6.2.21)

(ii) Consider a semi-explicit DAE with x := (xd, xa) ∈ Rnd+na and

F (t,x, x, u) :=

f (t, xd, xa, u) − xd

g(t, xd, xa, u)

∈ R

nd+na.

The corresponding adjoint equation (6.2.17) for λ := (λd, λa) ∈ Rnd+na reads as

λd(t)f xd[t] + λa(t)gxd

[t] + λd(t) = 0nd, (6.2.22)

λd(t)f xa[t] + λa(t)gxa

[t] = 0na . (6.2.23)

With S := (S d, S a) we findγ xf

+ λ(tf ) · F x[tf ]

· S (tf ) =

γ xd,f

− λd(tf )

S d(tf ) + γ xa,f S a(tf ). (6.2.24)

The task is to seek for a consistent value λ(tf ) = (λd(tf ), λa(tf )) for (6.2.22)-(6.2.23)such that the expression in (6.2.24) does not depend explicitly on S (tf ).

(a) Let g xa be non-singular a.e. in [t0, tf ]. Then, the DAE has index one and

λa(tf ) =

−λd(tf )

f xa[tf ](gxa

[tf ])−1

is consistent with the algebraic constraint (6.2.23) of the adjoint system for any λd(tf ).The expression (6.2.24) vanishes if

λd(tf ) = γ xd,f

,

γ xa,f = 0na .

Hence, with these settings, J (z) is given by (6.2.21).The latter assumption is not as restrictive as it seems as it fits well into the theoreticalinvestigations in Chapter 4. For, the algebraic component x a is an L∞-function andthus it makes no sense to allow a pointwise evaluation of xa at t0 or tf . Consequently,it is natural to prohibit γ depending explicitly on x a.

c 2006 by M. Gerdts

Page 166: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 166/243

158 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

(b) Let g xa ≡ 0 and g xd

f xa be non-singular a.e. in [t0, tf ]. Then, the DAE has index two

and

λa(tf ) = −λd(tf )

f xd

[tf ]f xa[tf ] − d

dtf xa

[tf ]

(gxd

[tf ]f xa[tf ])

−1

is consistent with the derivative of the algebraic constraint (6.2.23) for any λd(tf ).However, not any λf (tf ) is consistent with the constraint (6.2.23). We make theansatz

γ xd,f − λd(tf )

= ξ gxd[tf ]

with ξ ∈ Rna to be determined such that λd(tf ) is consistent with (6.2.23), i.e.

0na = λd(tf )f xa

[tf ] = γ xd,f f xa

[tf ] − ξ gxd[tf ]f xa

[tf ],

and thus

ξ

= γ xd,f f

xa[tf ] g

xd [tf ]f

xa [tf ]−1

.

Moreover, the sensitivity matrix S d satisfies a.e. the algebraic equation

gxd[t]S d(t) + gu[t]uM,z (t; w) = 0na

and thus the first term of the right handside in (6.2.24) computes to

γ xd,f

− λd(tf )

S d(tf ) = ξ gxd[tf ]S d(tf ) = −ξ gu[tf ]u

M,z (tf ; w).

Notice, that the expression on the right does not depend on S (tf ) anymore. Finally,if, as above, we assume γ xa,f

= 0na , then J (z) can be computed without computingS (tf ) according to

J (z) = −γ xd,f f xa

[tf ]

gxd[tf ]f xa

[tf ]−1

gu[tf ]uM,z (tf ; w)

+

γ x0 − λ(t0) · F x[t0]

· S (t0) + γ z +

tf

t0

λ(t) · F u[t] · uM,z (t; w)dt.

Example 6.2.5 In the special case of an ODE of type

F (t, x(t), x(t), uM

(t; w)) := f (t, x(t), uM

(t; w))−

x(t)

equations (6.2.17) and (6.2.20) reduce to

λ(t) = −λ(t)f x[t], λ(tf ) = γ xf

.

In Cao et al. [CLPS03] stability results for the forward and the adjoint system are derived. It isshown for explicit ODE’s and semi-explicit index-1 and index-2 Hessenberg DAE’s that stabilityis preserved for the adjoint DAE. For fully implicit index 0 and index 1 DAE’s the augmentedadjoint system in (6.2.25) below is stable, if the original DAE was stable, cf. Theorem 4.3 inCao et. [CLPS03].

c 2006 by M. Gerdts

Page 167: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 167/243

6.2. CALCULATION OF GRADIENTS 159

6.2.3.1 Numerical Adjoint Approximation by BDF

We discuss the application of the BDF method to the adjoint DAE. Firstly, the implicit DAE

F (t, x(t), x(t), uM (t; w)) = 0nx

is transformed to semi-explicit form:

0nx = x(t) − v(t),

0nx = F (t, x(t), v(t), uM (t; w)).

Using the function

H (t,x, x,v,u,λv , λx) :=

λv , λx

x − vF (t,x,v,u)

= λv (x − v) + λx F (t,x,v,u),

the adjoint DAE (6.2.17) for y = (x, v) is given by

H y[t] − d

dt

H y[t]

= 02nx .

Transposition yields

F x[t]

λx(t) − λv(t)

−λv(t) +

F x[t]

λx(t)

=

0nx

0nx

. (6.2.25)

This is a linear DAE in λ = (λv, λx). Employing the BDF discretization method for theintegration step tm+k−1

→tm+k and using the approximation

λv(tm+k) ≈ 1

h

ki=0

αiλv(tm+i)

yields the linear equation

F x[tm+k]

λx(tm+k) − 1

h

ki=0

αiλv(tm+i)

−λv(tm+k) +

F x[tm+k]

λx(tm+k)

=

0nx

0nx

for λx(tm+k) and λv(tm+k). The size of the equation can be reduced by introducing the secondequation

λv(tm+k) =

F x[tm+k]

λx(tm+k),

into the first equation. Herein, the relations

λv(tm+i) =

F x[tm+i]

λx(tm+i), i = 0, 1, . . . , k

are exploited. It remains to solve the linear equation

F x[tm+k] − αk

h F x[tm+k]

λx(tm+k) = 1

h

k−1

i=0

αi F x[tm+i] · λx(tm+i).

c 2006 by M. Gerdts

Page 168: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 168/243

Page 169: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 169/243

6.3. NUMERICAL EXAMPLE 161

with functions

L(v,h,C L) = q (v, h) F C L, ρ(h) = ρ0 exp(−βh),D(v,h,C L) = q (v, h) F C D(C L), R(h) = r0 + h,

C D(C L) = C D0 + k C L2, g(h) = g0(r0/R(h))2,q (v, h) = 1

2ρ(h)v2

and constants

F = 305, r0 = 6.371 · 106, C D0 = 0.017,k = 2, ρ0 = 1.249512, β = 1/6900,

g0 = 9.80665, ω = 7.27 · 10−5, m = 115000,

cf Chudej [Chu94]. Since the propulsion system is damaged, the mass m remains constant. Boxconstraints for the two control functions C L and µ are given by

0.01 ≤ C L ≤ 0.18326,

−π

2 ≤ µ ≤ π

2.

The initial values for the state correspond to a starting position above Bayreuth/Germany

v(0)γ (0)χ(0)h(0)Λ(0)Θ(0)

=

2150.54529000.15201817702.268927988933900.0000000.86515971020.1980948701

.

An additional restriction is given by the terminal condition

h(tf ) = 500.

The final time tf is assumed to be free and thus tf is an additional optimization variable.

The infinite dimensional optimal control problem is discretized by the reduced discretizationapproach. We used the fourth order classical Runge-Kutta method with fixed step-size fortime-integration of the ODE. The control is approximated by a continuous and piecewise linearfunction.

c 2006 by M. Gerdts

Page 170: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 170/243

162 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

0.860.88

0.90.92

0.94

0.040.08

0.120.16

0.2

0

10000

20000

30000

40000

50000

0.06

0.08

0.1

0.12

0.14

0.16

0.18

0.2

0 0.2 0.4 0.6 0.8 1

-0.1

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0 0.2 0.4 0.6 0.8 1

C L,app(t) µapp(t)

Λ

Θ

h

Figure 6.2: Numerical solution of the problem: 3D plot of the state (top) and approximateoptimal controls C L,app and µapp (bottom, normalized time scale) for N = 320 equidistantintervals.

Figure 6.2 shows the numerical solution for N = 320 equidistant intervals. The approximate

optimal final time for this highly nonlinear optimization problem is calculated to tf = 728.874966seconds and the approximate objective function value is −0.77581222.

Figure 6.3 summarizes computational results obtained for the sensitivity equation approachand the adjoint equation approach. While the sensitivity approach grows nonlinearly with thenumber N of equidistant time intervals in the control grid, the adjoint approach grows at a linearrate. Hence, in this case, the adjoint approach is more efficient than the sensitivity approach.This is the expected behavior since the number of constraints is significantly smaller than thenumber of optimization variables.

c 2006 by M. Gerdts

Page 171: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 171/243

Page 172: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 172/243

164 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

Assumption 6.4.2

(i) Let gy(t,x,y,u) be non-singular for every t, x, y,u.

(ii) For every t, x, y,u let rank (cu(t,x,y,u)) = nc.

(iii) Let the matrix g y−gu(cu)+cy be non-singular where (cu)+ denotes the pseudo-inverse of cu.

Problem 6.4.1 is discretized by application of the implicit Euler’s method on the (not necessarilyequidistant) grid GN in (6.1.1). The resulting finite dimensional nonlinear program reads asfollows.

Problem 6.4.3 (Discretized DAE optimal control problem)Find grid functions xN : GN → Rnx, xi := xN (ti), yN : GN → Rny , yi := yN (ti), uN : GN →Rnu, ui := uN (ti), such that

ϕ(x0, xN ) (6.4.8)

is minimized subject to the equations

f (ti, xi, yi, ui) − xi − xi−1

hi−1= 0nx , i = 1, . . . , N , (6.4.9)

g(ti, xi, yi, ui) = 0ny , i = 1, . . . , N , (6.4.10)

ψ(x0, xN ) = 0nψ, (6.4.11)

c(ti, xi, yi, ui) ≤ 0nc, i = 1, . . . , N , (6.4.12)

s(ti, xi) ≤ 0ns, i = 0, . . . , N , (6.4.13)

ui ∈ U , i = 1, . . . , N . (6.4.14)

A few remarks are in order. Equations (6.4.9)-(6.4.10) result from the implicit Euler’s methodapplied to the DAE (6.4.2)-(6.4.3). Notice, that we can dispense with the algebraic constraint(6.4.3) and the mixed control-state constraint (6.4.5) at t = t0 in the discretized problem if Assumption 6.4.2 is valid. This is true because Assumption 6.4.2 guarantees that the equationsg(t0, x0, y0, u0) = 0ny and c(t0, x0, y0, u0) = 0nc can be solved for y0 and u0 for arbitrary x0

provided that a solution exists at all. On the other hand, u0 and y0 would not enter the objectivefunction and thus it would be superfluous to impose the constraints g(t0, x0, y0, u0) = 0ny andc(t0, x0, y0, u0) ≤ 0nc in Problem 6.4.3. If Assumption 6.4.2 is not valid these constraints haveto be added. The following discrete local minimum principle can be modified accordingly in thiscase. Unfortunately, then the following interpretations concerning the relationship between themultipliers of Problem 6.4.1 and Problem 6.4.3 do not hold anymore. This lack of consistency

could be explained if one could show that the corresponding continuous minimum principle doesnot hold in the corresponding form. This question is currently under investigation.

Evaluation of the Fritz-John conditions in Theorem 3.6.2 yields the following necessary optimal-ity conditions.

Theorem 6.4.4 (Discrete Local Minimum Principle)

(i) Let the functions ϕ, f,g,c,s,ψ be continuously differentiable w.r.t. x, y, and u.

(ii) Let U ⊆ Rnu be a closed and convex set with non-empty interior.

(iii) Let (xN , yN , uN ) be a local minimum of Problem 6.4.3.

c 2006 by M. Gerdts

Page 173: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 173/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 165

Then there exist multipliers κ0 ∈ R, κ ∈ Rnψ , λf,N : GN → Rnx, λf,i := λf,N (ti), λg,N : GN →Rnx, λg,i := λg,N (ti), ζ N : GN → Rnc, ζ i := ζ N (ti), and ν N : GN → Rns, ν i := ν N (ti) such that the following conditions are satisfied:

(i) κ0 ≥ 0, (κ0, κ , λf,N , λg,N , ζ N , ν N ) = Θ,

(ii) Discrete adjoint equations:

For i = 1, . . . , N it holds

λf,i−1 = λf,i + hi−1Hx

ti, xi, yi, ui, λf,i−1, λg,i−1,

ζ ihi−1

+ ν i sx(ti, xi)

= λf,N +N

j=i

h j−1Hx

t j, x j , y j, u j , λf,j−1, λg,j−1,

ζ jh j−1

+N

j=i

ν j sx(t j, x j), (6.4.15)

0ny = Hy

ti, xi, yi, ui, λf,i−1, λg,i−1,

ζ ihi−1

. (6.4.16)

(iii) Discrete transversality conditions:

λf,0 = −

κ0ϕx0(x0, xN ) + κψx0(x0, xN ) + ν 0 sx(t0, x0)

, (6.4.17)

λ

f,N

= κ0ϕ

xN

(x0, xN ) + κψ

xN

(x0, xN ). (6.4.18)

(iv) Discrete optimality conditions:For all i = 1, . . . , N and all u ∈ U it holds

Hu

ti, xi, yi, ui, λf,i−1, λg,i−1,

ζ ihi−1

(u − ui) ≥ 0. (6.4.19)

(v) Discrete complementarity conditions:It holds

ζ i ≥ 0nc , i = 1, . . . , N , (6.4.20)

ν i ≥ 0ns , i = 0, . . . , N , (6.4.21)

ζ i c(ti, xi, yi, ui) = 0, i = 1, . . . , N , (6.4.22)

ν i s(ti, xi) = 0, i = 0, . . . , N . (6.4.23)

Proof. Let κ0 ∈ R, x = (x0, . . . , xN ) ∈ Rnx(N +1), y = (y1, . . . , yN )

∈ RnyN , u =(u1, . . . , uN )

∈ RnuN , λf = (λf,0, . . . , λf,N −1) ∈ RnxN , λg = (λg,0, . . . , λg,N −1) ∈ RnyN ,ζ = (ζ 1, . . . , ζ N )

∈ RncN , ν = (ν 0, . . . , ν N )

∈ Rns(N +1), κ

∈ Rnψ . The Lagrange function of

c 2006 by M. Gerdts

Page 174: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 174/243

166 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

Problem 6.4.3 is given by

L(x,y,u, λf , λg, ζ , ν , κ , κ0) := κ0ϕ(x0, xN ) + κψ(x0, xN )

+N

i=1

λf,i−1f (ti, xi, yi, ui)

− xi − xi−1

hi−1 +

N i=1

λg,i−1g(ti, xi, yi, ui) +N

i=1

ζ i c(ti, xi, yi, ui) +N

i=0

ν i s(ti, xi)

= κ0ϕ(x0, xN ) + κψ(x0, xN )

+N

i=1

H(t, xi, yi, ui, λf,i−1, λg,i−1, ζ i) +N

i=1

λf,i−1

xi−1 − xi

hi−1

+N

i=0

ν i s(ti, xi).

Application of Theorem 3.6.2 results in the following equations:

1. Lui

(u − ui) ≥ 0 ∀u ∈ U : For i = 1, . . . , N it holds

Hu(ti, xi, yi, ui, λf,i−1, λg,i−1, ζ i)(u − ui) ≥ 0 ∀u ∈ U .

2. Lyi

= 0: For i = 1, . . . , N it holds

Hy(ti, xi, yi, ui, λf,i−1, λg,i−1, ζ i) = 0ny

3. Lxi

= 0: For i = 0 it holdsκ0ϕx0(x0, xN ) + κψ

x0(x0, xN ) + ν 0 sx(t0, x0)

+ 1

h0λf,0 = 0nx .

For i = 1, . . . , N − 1 it holds

Hx(ti, xi, yi, ui, λf,i−1, λg,i−1, ζ i) − 1

hi−1λf,i−1 +

1

hiλf,i + ν i sx(ti, xi) = 0nx

For i = N it holds

κ0ϕxN (x0, xN ) + κψ

xN (x0, xN ) + ν N s

x(tN , xN )

+ Hx(tN , xN , yN , uN , λf,N −1, λg,N −1, ζ N ) − 1hN −1

λf,N −1 = 0nx

With the definitions

λf,i := 1

hiλf,i , λg,i :=

1

hiλg,i, i = 0, . . . , N − 1,

andλf,N := κ0ϕxN

(x0, xN ) + κψxN

(x0, xN ).

we obtain the discrete optimality conditions, the discrete adjoint equations, and the discretetransversality conditions.

c 2006 by M. Gerdts

Page 175: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 175/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 167

Remark 6.4.5 The local minimum principle for ODE optimal control problems can be extended to a global minimum principle. The question arises whether a similar result holds for the dis-cretized optimal control problem. This is not the case in general. However, an approximate minimum principle holds, cf. Mordukhovich [Mor88]. With additional convexity-like conditions a discrete minimum principle holds as well, cf. Ioffe and Tihomirov [IT79], Section 6.4, p. 277.

We compare the necessary optimality conditions in Theorem 6.4.4 with those in Theorems 4.2.4and 4.2.5. Our intention is to provide an interpretation of the ‘discrete’ multipliers in Theo-rem 6.4.4 and put them into relation to the ‘continuous’ multipliers in Theorem 4.2.4. Followingthis interpretation it becomes possible to construct estimates of the continuous multipliers byuse of the discrete multipliers only. Of course, the discrete multipliers can be computed numeri-cally by solving the discretized optimal control problem 6.4.3 by SQP. Unfortunately, we are notable to present rigorous convergence proofs which would justify the following interpretations.This is current research. Nevertheless, we will summarize some available convergence results forODE optimal control problems in the next section.

6.4.1 Problems with Pure State Constraints

In line with the situation of Theorem 4.2.4 we discuss the Problem 6.4.1 without mixed control-state constraints, i.e. nc = 0 and c ≡ 0.The discrete optimality conditions (6.4.19) and the discrete adjoint equations (6.4.16) are easilybeing recognized as pointwise discretizations of the optimality condition (4.2.12) and the alge-braic equation in (4.2.9), respectively, provided we assume for all i (actually there will be anexceptional interpretation for λf,0 later on):

xi ≈ x(ti), yi ≈ y(ti), ui ≈ u(ti), λf,i ≈ λf (ti), λg,i ≈ λg(ti).

Comparing the discrete transversality condition (6.4.18)

λf,N = κ0ϕxN (x0, xN ) + κψ

xf (x0, xN )

and the transversality condition (4.2.11)

λf (tf ) = l0ϕxf

(x(t0), x(tf )) + σψxf

(x(t0), x(tf ))

it is natural to presumeκ0 ≈ l0, κ ≈ σ.

The adjoint equation (4.2.8) for t ∈ [t0, tf ] reads as

λf (t) = λf (tf ) + tf

tHx(τ, x(τ ), y(τ ), u(τ ), λf (τ ), λg(τ ))dτ + tf

tsx(τ, x(τ ))dµ(τ ),

where we used the abbreviation tf

tsx(τ, x(τ ))dµ(τ ) :=

nsi=1

tf

tsi,x[τ ]dµi(τ ).

The discrete adjoint equation (6.4.15) for i = 1, . . . , N is given by

λf,i−1 = λf,N +N

j=i

h j−1Hx (t j, x j , y j, u j , λf,j−1, λg,j−1) +

N

j=i

sx(t j , x j )ν j .

c 2006 by M. Gerdts

Page 176: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 176/243

168 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

The first sum is easily being recognized to be a Riemann sum on GN and thus

N j=i

h j−1Hx (t j, x j , y j, u j , λf,j−1, λg,j−1) ≈

tf

ti−1

Hx(τ, x(τ ), y(τ ), u(τ ), λf (τ ), λg(τ ))dτ

(6.4.24)for i = 1, . . . , N . This observation encourages us to expect that for i = 1, . . . , N

N j=i

sx(t j, x j )ν j ≈ tf

ti−1

sx(τ, x(τ ))dµ(τ ). (6.4.25)

In order to interpret the approximation in (6.4.25) we have to recall the definition of a Riemann-Stieltjes integral, cf. Section 2.4. The Riemann-Stieltjes integral in (6.4.25) is defined to be thelimit of the sum

m j=i

sx(ξ j , x(ξ j )) (µ(t j) − µ(t j−1))

where ti−1 < ti < . . . < tm = tf is an arbitrary partition of [ti−1, tf ] and ξ j ∈ [t j−1, t j] arearbitrary points. If we chose the particular grid GN and the points ξ j := t j we obtain theapproximation tf

ti−1

sx(τ, x(τ ))dµ(τ ) ≈N

j=i

sx(t j , x(t j )) (µ(t j) − µ(t j−1)) .

Together with (6.4.25) we draw the conclusion that the relationship of the discrete multipliersν i and the continuous counterpart µ must be given by

ν i ≈ µ(ti) − µ(ti−1), i = 1, . . . , N .

Now, we address the remaining transversality condition (4.2.10):

λf (t0) = −

l0ϕx0(x(t0), x(tf )) + σψx0(x(t0), x(tf ))

and the discrete counterpart (6.4.17)

λf,0 = −

κ0ϕx0(x0, xN ) + κψx0(x0, xN ) + ν 0 sx(t0, x0)

.

These two conditions can be brought into accordance as follows. Recall, that the multiplier µis normalized, i.e. µ(t0) = 0ns and µ is continuous from the right in the open interval ( t0, tf ).Hence, µ may jump at t0 with 0 ≤ µ(t+0 ) − µ(t0) = µ(t+0 ). Similarly, λf may jump at t0, too.

Similar to the derivation of the jump conditions (4.2.21) it follows

λf (t+0 ) − λf (t0) = limε↓0

λf (t0 + ε) − λf (t0) = −sx(t0, x(t0))(µ(t+0 ) − µ(t0))

and thus

λf (t+0 ) = λf (t0) − sx(t0, x(t0))(µ(t+0 ) − µ(t0))

= −

l0ϕx0(x(t0), x(tf )) + σψx0(x(t0), x(tf )) + sx(t0, x(t0))(µ(t+0 ) − µ(t0))

.

A comparison with (6.4.17) shows that both are in accordance if ν 0 is interpreted as the jumpheight of µ at t = t0, i.e.

ν 0 ≈

µ(t+

0

)−

µ(t0)

c 2006 by M. Gerdts

Page 177: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 177/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 169

and if λf,0 is interpreted as the right-sided limit of λ f at t0, i.e.

λf,0 ≈ λf (t+0 ).

As an approximation of the value λ f (t0) we then use the value

κ0ϕx0(x0, xN ) + κψx0(x0, xN )

.

The above interpretations cope well with the complementarity conditions if we recall that ν idenotes the multiplier for the constraint s(ti, xi) ≤ 0ns . Then, the complementarity conditionsyield

0ns ≤ ν i ≈ µ(ti) − µ(ti−1)

which reflects the fact that µ is non-decreasing. The condition ν i s(ti, xi) = 0 implies

s(ti, xi) < 0ns ⇒ 0ns = ν i ≈ µ(ti) − µ(ti−1),

which reflects that µ is constant on inactive arcs.

Remark 6.4.6

• The above interpretations remain valid for problems with additional mixed control-state constraints and U = Rnu, cf. Theorem 4.2.5. A comparison of the respective necessary conditions in Theorem 4.2.5 and Theorem 6.4.4 yields the additional relationship

ζ ihi−1 ≈ η(ti), i = 1, . . . , N

for the multiplier η. The complementarity conditions for ζ i are discrete versions of the complementarity condition in Theorem 4.2.5.

• Instead of implicit Euler’s method more general Runge-Kutta methods can be investi-gated provided that these are suitable for DAE optimal control problems. For ODE op-timal control problems and explicit Euler’s method the analogous results can be found in Gerdts [Ger05b].

6.4.2 Example

The subsequent example shows that the above interpretations are meaningful. Since the problemis an ODE optimal control problem we use explicit Euler’s method for discretization instead of implicit Euler’s method. The above interpretations can be adapted accordingly. Though theexample at a first glance is very simple it has the nice feature that one of the two state constraintsbecomes active only at the final time point. This causes the corresponding multiplier to jumponly at the final time point. Correspondingly, the adjoint also jumps. It turns out that the aboveinterpretations allow to construct approximations for the multipliers that reflect this behaviorfor the numerical solution.

Consider a system of two water boxes, where x(t) and y(t) denote the volume of water in thetwo boxes and u(t) and v(t) denote the outflow rate of water for the respective boxes at time t,cf. Figure 6.4.

c 2006 by M. Gerdts

Page 178: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 178/243

170 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

x(t)

u(t)

y(t)

v(t)

Figure 6.4: System of two water boxes with controllable outflow rates.

Assume, that the outflow rates and the volumes are restricted by u(t), v(t) ∈ [0, 1] and x(t) ≥ 0,y(t) ≥ 0, respectively. We consider the following linear optimal control problem.

Problem 6.4.7 Minimize

− 100

(10 − t)u(t) + tv(t)dt

subject to the ODE

x(t) = −u(t),y(t) = u(t) − v(t),

the initial conditions x(0) = 4, y(0) = 4,

the state constraints x(t) ≥ 0, y(t) ≥ 0 in [0, 10],

and the control constraints

u(t) ∈ [0, 1], v(t) ∈ [0, 1] a.e. in [0, 10].

Evaluation of the local minimum principle 4.2.4 resp. 4.1.10 yields the following result.

Theorem 6.4.8 The following functions satisfy the local minimum principle 4.2.4 for the linear optimal control problem 6.4.7 and hence are optimal since the problem is convex.The optimal control variables are given by

u(t) =

1, if 0 ≤ t < 4,0, if 4 ≤ t ≤ 10,

, v(t) =

0, if 0 ≤ t < 2,1, if 2 ≤ t ≤ 10.

The optimal state variables are given by

x(t) =

4 − t, if 0 ≤ t < 4,0, if 4

≤ t

≤ 10,

, y(t) =

4 + t, if 0 ≤ t < 2,6, if 2 ≤ t < 4,10

−t, if 4

≤ t ≤

10.

c 2006 by M. Gerdts

Page 179: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 179/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 171

The adjoints are given by

λx(t) = µx(t) − µx(10), λy(t) = µy(t) − µy(10) =

−2, if 0 ≤ t < 10,0, if t = 10,

where the multiplier µx is non-decreasing and satisfies µx(t) = 0 for t ∈ [0, 4) and 104

dµx(τ ) = µx(10) − µx(4) = 8, µx(10) − µx(t) > 12 − t, t ∈ (4, 10).

The multiplier µy is given by

µy(t) =

0, if 0 ≤ t < 10,2, if t = 10.

Proof. We apply Theorem 4.2.4. The Hamilton function for Problem 6.4.7 is given by

H(t,x,y,u,v,λx, λy, l0) = −l0 ((10 − t)u + tv) − λxu + λy(u − v).

Let (x, y, u, v) be a local minimum of Problem 6.4.7. Then, by theorem 4.2.4 there exist mul-tipliers l0 ≥ 0, σ = (σx, σy) ∈ R2, λ = (λx, λy) ∈ BV ([t0, tf ],R

2), and µ = (µx, µy) ∈N BV ([t0, tf ],R

2) with

Θ = (l0,σ,λ,µ),

λx(tf ) = 0,

λy(tf ) = 0,

λx(t0) = −σx,

λy(t0) =

−σy,

λx(t) = λx(10) + 10

tH

x[τ ]dτ − 10

tdµx(τ ) = µx(t) − µx(10),

λy(t) = λy(10) +

10t

Hy[τ ]dτ −

10t

dµy(τ ) = µy(t) − µy(10).

Furthermore, a.e. in [0, 10] it holds

0 ≤ Hu[t](u − u(t)) = (−l0(10 − t) − λx(t) + λy(t))(u − u(t))

= (−l0(10 − t) − µx(t) + µx(10) + µy(t) − µy(10))(u − u(t)) ∀u ∈ [0, 1],

0 ≤ Hv[t](v − v(t)) = (−l0t − λy(t))(v − v(t))

= (−l0t − µy(t) + µy(10))(v − v(t)) ∀v ∈ [0, 1].

Finally, µx, µy are non-decreasing functions with µx(0) = µy(0) = 0. µx and µy are constant onintervals with measure greater than zero and x(t) > 0 and y (t) > 0, respectively.First, notice that the case l0 = 0 can be excluded, since the Mangasarian-Fromowitz conditionsare easily being checked to be satisfied for this problem. Hence, we may set l 0 = 1.From the optimality conditions we conclude

u(t) =

0, if λy(t) − λx(t) > 10 − t,1, if λy(t) − λx(t) < 10 − t,

v(t) = 1, if − λy(t) < t,0, if

−λ

y(t) > t.

c 2006 by M. Gerdts

Page 180: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 180/243

172 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

The monotonicity properties of µx and µy imply that λx(t) = µx(t) − µy(10) ≤ 0 and λy(t) =µy(t) − µy(10) ≤ 0 are non-decreasing functions. Since the function t is strictly increasingand −λy(t) ≥ 0 is monotonically decreasing, there exists at most one point t ∈ [0, 10] with−λy(t) = t. Hence, the control v is a bang-bang control with at most one switching point t.This defines the structure of the control v.The first differential equation yields

x(t) = 4 − t

0u(τ )dτ, 0 ≤ t ≤ 10.

Let us assume for a moment, that x(t) > 0 holds for all t ∈ [0, 10], i.e. the state constraintx(t) ≥ 0 is inactive. Then, µx(t) ≡ 0 and λx(t) = µx(t) − µx(10) ≡ 0. The optimality conditionyields u(t) ≡ 1, because λy(t) − λx(t) = λy(t) ≤ 0 ≤ 10 − t. But this contradicts x(t) > 0 for allt ∈ [0, 10]. Hence, there exists a first point t ∈ (0, 10] with x(t) = 0. Once x(t) = 0 is fulfilled,it holds x(t) = 0 and u(t) = 0 for all t ∈ [ t, 10] because of

0 ≤ x(t) = x(t) − t

tu(τ )dτ = − t

tu(τ )dτ ≤ 0.

Due to the optimality conditions this implies λ y(t) − λx(t) ≥ 10 − t in (t, 10].Since t is the first point satisfying x(t) = 0 it holds x(t) > 0 in [0, t) and thus µx(t) ≡ 0respectively λx(t) ≡ −µx(10) in [0, t). Hence, λy(t) − λx(t) is non-decreasing in [0, t). Since10 − t is strictly decreasing, there is at most one point t1 ∈ [0, t) with λy(t1) − λx(t1) = 10 − t1.Assume t1 < t. Then, necessarily, u(t) = 0 in (t1, t) and thus u(t) = 0 in (t1, 10] and x(t) ≡ x(t1)in [t1, 10]. Then, either x(t1) > 0, which contradicts the existence of t, or x(t1) = 0, whichcontradicts the minimality of t. Consequently, there is no point t1 in [0, t) with λy(t1)−λx(t1) =10 − t1. Therefore, either λy(t) − λx(t) > 10 − t in [0, t), which implies u(t) = 0 and contradictsthe existence of t, or λy(t)

−λx(t) < 10

−t in [0, t), which implies u(t) = 1 in [0, t). Summarizing,

these considerations yield

t = 4,

u(t) =

1, if t ∈ [0, 4),0, if t ∈ [4, 10],

x(t) =

4 − t, if t ∈ [0, 4),0, if t ∈ [4, 10],

µx(t) = 0, t ∈ [0, 4),

λx(t) = µx(t) − µx(10) = −µx(10), t ∈ [0, 4),

λy(t)

−λx(t) = 10

−t = 6.

Now, we have to determine the switching point t of v. It turns out that t = 2 is the only possiblechoice. All other cases (t does not occur in [0, 10], t = 2) will lead to contradictions. Hence, wehave

t = 2, v(t) =

0, if t ∈ [0, 2),1, if t ∈ [2, 10].

Using these controls, it is easy to check that y becomes active exactly at t = 10. Hence,

µy(t) =

0, in [0, 10),µy(10) ≥ 0, if t = 10,

λy(t) = µy(t) − µy(10) =

−µy(10), in [0, 10),0, if t = 10.

On the other hand we know already t = −

λy(t) and hence µy(10) = 2.

c 2006 by M. Gerdts

Page 181: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 181/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 173

Moreover, µx(t) = 0 in [0, 4) and hence

λx(t) = µx(t) − µx(10) =

−µx(10), if 0 ≤ t < 4,≤ 0, if 4 ≤ t < 10,0, if t = 10.

On the other hand we know already 6 = λy(t) − λx(t) = −2 − λx(t) and hence λx(t) = µx(t) −µx(10) = − 10

t dµx(τ ) = −8. Finally, since u(t) ≡ 0 in [4, 10] it follows λy(t) − λx(t) =−2 − λx(t) > 10 − t in (4, 10), i.e. µx(10) − µx(t) > 12 − t in (0, 4).

In the sequel, the discretization of Problem 6.4.7 by explicit Euler’s method is investigated indetail. Application of explicit Euler’s method with constant step size h = (tf − t0)/N andequidistant grid points ti = t0 + ih, i = 0, . . . , N to Problem 6.4.7 leads to

Problem 6.4.9 (Discretized Problem)Minimize

−hN −1i=0

(10 − ti)ui + tivi

subject to the constraints

x0 = 4,xi+1 = xi − hui, i = 0, 1, . . . , N − 1,

y0 = 4,yi+1 = yi + hui − hvi, i = 0, 1, . . . , N − 1,

xi ≥ 0, i = 0, 1, . . . , N ,yi ≥ 0, i = 0, 1, . . . , N ,ui

∈ [0, 1], i = 0, 1, . . . , N

−1,

vi ∈ [0, 1], i = 0, 1, . . . , N − 1.

The Lagrange function of Problem 6.4.9 with x = (x1, . . . , xN ), y = (y1, . . . , yN )

, u =(u0, . . . , uN −1), v = (v0, . . . , vN −1), λx = (λx

1 , . . . , λxN )

, λy = (λy1, . . . , λy

N ), µx = (µx

1 , . . . , µxN )

,µy = (µy

1, . . . , µyN )

, is given by

L(x,y,u,v,λx, λy, µx, µy) = −hN −1i=0

[(10 − ti)ui + tivi]

+N −1

i=0

λxi+1(xi − hui − xi+1)

+N −1i=0

λyi+1(yi + hui − hvi − yi+1)

+N

i=0

µxi (−xi) +

N i=0

µyi (−yi).

Notice, that x0 = 4 and y0 = 4 are not considered as constraints in Problem 6.4.9.We intend to evaluate the Fritz-John conditions in Theorem 3.6.2 for Problem 6.4.9. Notice,that we can choose l0 = 1, since Problem 6.4.9 is linear. Theorem 3.6.2 with l0 = 1 yields thefollowing necessary (and in this case also sufficient) optimality conditions for an optimal solutionx, y, u, v with multipliers λx, λy , µx, µy:

c 2006 by M. Gerdts

Page 182: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 182/243

174 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

(i) For i = 1, . . . , N it holds 0 = Lxi

(x, y, u, v, λx, λy, µx, µy), i.e.

0 = λxi+1 − λx

i − µxi , i = 1, . . . , N − 1,

0 = λxN − µx

N .

Recursive evaluation leads to

λxi = −

N j=i

µx j , i = 1, . . . , N .

(ii) For i = 1, . . . , N it holds 0 = Lyi

(x, y, u, v, λx, λy, µx, µy), i.e.

0 = λyi+1 − λy

i − µyi , i = 1, . . . , N − 1,

0 = λyN − µy

N .

Recursive evaluation leads to

λyi = −

N j=i

µy j , i = 1, . . . , N .

(iii) For i = 0, . . . , N it holds L ui

(x, y, u, v, λx, λy, µx, µy)(u − ui) ≥ 0 for all u ∈ [0, 1], i.e.−h(10 − ti) − hλxi+1 + hλy

i+1

(u − ui) ≥ 0,

respectively

−h(10 − ti) + h

N j=i+1

µx j − µy ju ≥ −h(10 − ti) + h

N j=i+1

µx j − µy j ui

for all u ∈ [0, 1]. This implies

ui =

1, if − (10 − ti) +N

j=i+1(µx j − µy

j ) < 0,

0, if − (10 − ti) +N

j=i+1(µx j − µy

j ) > 0,

undefined, otherwise.

(6.4.26)

(iv) For i = 0, . . . , N it holds L vi

(x, y, u, v, λx, λy, µx, µy)(v − vi) ≥ 0 for all v ∈ [0, 1], i.e.

−hti − hλyi+1 (v − vi) ≥ 0,

respectively −hti + h

N j=i+1

µy j

v ≥

−hti + h

N j=i+1

µy j

vi

for all v ∈ [0, 1]. This implies

vi =

1, if − ti +N

j=i+1 µy j < 0,

0, if − ti +

N j=i+1 µy

j > 0,

undefined, otherwise.

(6.4.27)

c 2006 by M. Gerdts

Page 183: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 183/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 175

(v) It holds µxi xi = 0, µx

i ≥ 0, i = 0, . . . , N , and µyi yi = 0, µy

i ≥ 0, i = 0, . . . , N .

We will check that

ui = 1, for i = 0, . . . , m

−1,

4/h − m, for i = m,0, for i = m + 1, . . . , N − 1,

(6.4.28)

vi =

0, for i = 0, . . . , q − 1,q + 1 − 2/h, for i = q,1, for i = q + 1, . . . , N − 1,

(6.4.29)

with

m =

4

h

, q =

2

h

,

is an optimal control for the discretized problem 6.4.9. By application of these controls we find

xi = 4 − hi−1

j=0

u j = 4 − ih > 0, i = 0, . . . , m ,

xm+1 = xm − hum = 4 − mh − h(4/h − m) = 0,

xi = 0, i = m + 2, . . . , N .

Similarly, we have

yi = 4 + hi−1

j=0

(u j − v j) = 4 + ih, i = 0, . . . , q ,

yq+1 = yq + h(uq − vq) = 4 + qh + h(1 − (q + 1 − 2/h)) = 6,yi = 6, i = q + 2, . . . , m ,

ym+1 = ym + h(um − vm) = 6 + h(4/h − m − 1) = 10 − h(m + 1),

yi = ym+1 + hi−1

j=m+1

(u j − v j )

= 10 − h(m + 1) − h(i − 1 − m) = 10 − ih, i = m + 2, . . . , N .

Notice, that yi > 0 for i = 0, . . . , N − 1 and yN = 0. Hence, according the complementarityconditions (v) the multipliers µy

i must satisfy µyi = 0 for i = 1, . . . , N − 1. With (6.4.27) and

(6.4.29) necessarily it holds

µyN ∈ (tq−1, tq).

Taking the limit h → 0 we find tq = qh = 2/hh → 2 and tq−1 → 2 and thus µyN → 2.

According to the complementarity conditions (v) the multipliers µxi must satisfy µx

i = 0 fori = 1, . . . , m. With (6.4.26) and (6.4.28) necessarily it holds

−(10 − tm−1) +N

j=m

µx

j − µy j

< 0,

−(10 − tm) +N

j=m+1µx

j − µy j

> 0,

c 2006 by M. Gerdts

Page 184: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 184/243

176 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

and thus

µyN + (10 − tm) <

N j=m

µx j < µy

N + (10 − tm−1).

Taking the limit h → 0 yields tm = mh = 4/hh → 4 and

N j=m

µx j → 8.

Summarizing, the above considerations showed that (6.4.28), (6.4.29), and the resulting discretestate variables are optimal solutions of the discretized problem provided the multipliers µ x andµy are chosen accordingly. Furthermore, the switching points tq and tm converge at a linearrate to the switching points of the continuous problem 6.4.7. The discrete states (viewed as

continuous, piecewise linear functions) converge for the norm · 1,∞, whereas the discretecontrols (viewed as piecewise constant functions) do not converge to the continuous controls forthe norm · ∞. However, the discrete controls do converge for the norm · 1. Due to thenon-uniqueness of the continuous and discrete multipliers it is hard to observe convergence forthese quantities.

Figure 6.5 shows the numerical solution for N = 999 and piecewise constant control approxima-tion. The numerical solution was computed by the software package Sodas, cf. Gerdts [Ger01b].

-0.5

0

0.5

1

1.5

2

2.5

3

3.5

4

0 2 4 6 8 10

0

1

2

3

4

5

6

0 2 4 6 8 10

0

0.2

0.4

0.6

0.8

1

0 2 4 6 8 10

0

0.2

0.4

0.6

0.8

1

0 2 4 6 8 10

x ( t )

y ( t )

u ( t )

v ( t )

Figure 6.5: Optimal approximate state and control for N = 999 grid points and piecewiseconstant control approximation.

c 2006 by M. Gerdts

Page 185: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 185/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 177

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

0

1

2

3

4

5

6

7

8

0 2 4 6 8 10

N = 9 N = 19

N = 39 N = 79

N = 159 N = 319

N = 639 N = 999

Figure 6.6: Lagrange multiplier µx for the discretized state constraint xi ≥ 0 in Problem 6.4.9for N = 9, 19, 39, 79, 159, 319, 639, 999.

c 2006 by M. Gerdts

Page 186: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 186/243

Page 187: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 187/243

6.4. DISCRETE MINIMUM PRINCIPLE AND APPROXIMATION OF ADJOINTS 179

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

-8

-7

-6

-5

-4

-3

-2

-1

0

0 2 4 6 8 10

N = 9 N = 19

N = 39 N = 79

N = 159 N = 319

N = 639 N = 999

Figure 6.8: Adjoint estimation λx for the discretized differential equation in Problem 6.4.9 forN = 9, 19, 39, 79, 159, 319, 639, 999.

c 2006 by M. Gerdts

Page 188: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 188/243

180 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

-2

-1.5

-1

-0.5

0

0 2 4 6 8 10

N = 9 N = 19

N = 39 N = 79

N = 159 N = 319

N = 639 N = 999

Figure 6.9: Adjoint estimation λy for the discretized differential equation in Problem 6.4.9 forN = 9, 19, 39, 79, 159, 319, 639, 999.

c 2006 by M. Gerdts

Page 189: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 189/243

6.5. CONVERGENCE 181

6.5 Convergence

The convergence of discretized optimal control problems is a current field of research. Onlyfew results are available for ODE optimal control problems. We summarize the existing resultswithout proving them. All results assume that the optimal control is at least continuous. Up to

now, the author is not aware of any results considering discontinuous optimal controls or ODEoptimal control problems subject to pure state constraints. Hence, a lot of research has to bedone in this direction!

6.5.1 Convergence of the Euler Discretization

The proof of convergence for the full Euler discretization can be found in Malanowski etal. [MBM97]. More specifically, the authors investigate the optimal control problem

min ϕ(x(tf )) +

tf

0f 0(x(t), u(t))dt

s.t. x(t) − f (x(t), u(t)) = 0nx ,

x(0) − ξ = 0nx , ψ(x(tf )) = 0nψ ,c(x(t), u(t)) ≤ 0nc ,

and its discretization by Euler’s method:

min ϕ(xN ) + hN −1i=0

f 0(xi, ui)

s.t. xi+1 − xi

h − f (xi, ui) = 0nx , i = 0, . . . , N − 1,

x0 − ξ = 0nx , ψ(xN ) = 0nψ,

c(xi, ui) ≤ 0nc, i = 0, . . . , N − 1.

We only summarize the assumptions needed to prove a convergence result. As usual H denotesthe augmented Hamilton function.

Assumption 6.5.1

(i) f 0,f ,c,ϕ,ψ are differentiable with locally Lipschitz continuous derivatives.

(ii) There exists a local solution (x, u) ∈ C 1([0, tf ],Rnx)×C ([0, tf ],R

nu) of the optimal control problem.

(iii) Uniform rank condition for c:Let J (t) := i ∈ 1, . . . , nc | ci(x(t), u(t)) = 0 denote the index set of active mixed control-state constraints at t and cJ (t)[t] the active constraints at t. Let there exists a constant α > 0 with

cJ (t),u[t]d ≥ αd ∀d ∈ R|J (t)| a.e. in [0, tf ].

(iv) Surjectivity of the linearized equality constraints:Let the boundary value problem

y(t)−

A(t)y(t)−

B(t)v(t) = 0nx , y(0) = 0nx , ψ

xf

(x(tf ))y(tf ) = h

c 2006 by M. Gerdts

Page 190: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 190/243

182 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

possess a solution for any h ∈ Rnψ , where

A(t) = f x[t] − f u[t]cJ (t),u[t]

cJ (t),u[t]cJ (t),u[t]−1

cJ (t),x[t],

B(t) = f

u

[t]I nu

−c

J (t),u

[t] c

J (t),u

[t]c

J (t),u

[t]−1c

J (t),x

[t] .

(v) Coercivity:Define J +(t) := i ∈ J (t) | ηi(t) > 0, where η denotes the multiplier for the mixed control-state constraint c.

Let there exist β > 0 such that d H

uu[t]d ≥ β d2

holds for all d ∈ Rnu with cJ +(t),u[t]d = 0|J +(t)|.

(vi) Riccati equation:

Let the Riccati equation

Q(t) = −Q(t)f x[t] − f x[t]Q(t) − Hxx[t]

+

H

ux[t]cJ +(t),x[t]

+ Q(t)

f u[t]

Θ

Huu[t] c

J +(t),u[t]

cJ +(t),u

[t] Θ

·

·

f u[t]

Θ

Q(t) +

H

ux[t]cJ +(t),x[t]

possess a bounded solution Q on [0, tf ] that satisfies the rank condition

d (Γ − Q(tf )) d ≥ 0 ∀d ∈ Rnx : ψxf (x(tf ))d = 0nψ ,

where

Γ :=

ϕ(x(tf )) + σf ψ(x(tf ))

xx.

All function evaluations are at the optimal solution (x, u).

Theorem 6.5.2 Let the assumptions (i)-(vi) in Assumption 6.5.1 be satisfied. Then, for suf- ficiently small step-sizes h > 0 there exist a locally unique KKT point (xh, uh, λh, ζ h, κ0, κf ) of the discretized problem satisfying

max

xh

− x

1,∞,

uh

− u

∞,

λh

−λ

1,∞,

κ0

−σ0

,

κf

−σf

,

ηh

−η

=

O(h),

where λh denotes the discrete adjoint, ηh the discrete multiplier for the mixed control-state constraint, κ0 the discrete multiplier for the initial condition, and κf the discrete multiplier for the final condition.

Proof. The long and complicated proof can be found in Malanowski et al. [MBM97].

Remark 6.5.3

• The assumptions (v) and (vi) together are sufficient for local optimality of (x, u).

• Similar convergence results can be found in Dontchev et al. [DHV00, DHM00].

c 2006 by M. Gerdts

Page 191: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 191/243

6.5. CONVERGENCE 183

6.5.2 Higher Order of Convergence for Runge-Kutta Discretizations

Hager [Hag00] investigates optimal control problems of type

min ϕ(x(1))

s.t. x(t) = f (x(t), u(t)), x(0) = ξ, t ∈ [0, 1],

and their discretization by s-staged Runge-Kutta methods with fixed step-size h = 1/N :

min ϕ(xN )

s.t. xk+1 − xk

h =

si=1

bif (ηi, uki), k = 0, . . . , N − 1,

ηi = xk + hs

j=1

aij f (η j, ukj ), i = 1, . . . , s ,

x0 = ξ.

Notice, that for each function evaluation of f an independent optimization variable u ki is intro-duced. Hence, there is no coupling by, e.g., piecewise constant or linear interpolation.

The main result of Hager’s article is a convergence result for the discrete solutions assuming thefollowing:

Assumption 6.5.4

(i) Smoothness:The optimal control problem possesses a solution (x, u) ∈ W p,∞([0, 1],Rnx)×W p−1,∞([0, 1],Rnu)with p ≥ 2.

The first p derivatives of f and ϕ are supposed to be locally Lipschitz continuous in some neighborhood of (x, u).

(ii) Coercivity:There exists some α > 0 with

B(x, u) ≥ αu22 ∀(x, u) ∈ M

where

B(x, u) = 1

2 x(1)V x(1) +

1

0x(t)Q(t)x(t) + 2x(t)S (t)u(t) + u(t)R(t)u(t)dt

and

A(t) := f x(x(t), u(t)), B(t) := f u(x(t), u(t)),

V := ϕ(x(1)), Q(t) := Hxx(x(t), u(t), λ(t)),

R(t) := Huu(x(t), u(t), λ(t)), S (t) := H

xu(x(t), u(t), λ(t)),

and

M =

(x, u) ∈ W 1,2([0, 1],Rnx) × L2([0, 1],Rnu)

x = Ax + Bu, x(0) = 0

.

c 2006 by M. Gerdts

Page 192: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 192/243

184 CHAPTER 6. DISCRETIZATION OF OPTIMAL CONTROL PROBLEMS

The smoothness property in (i) implies that the optimal control u is at least continuous.It turns out that the well-known order conditions for Runge-Kutta methods for initial valueproblems are not enough to ensure higher order convergence for discretized optimal controlproblems. To ensure higher order convergence for the discretized optimal control problem theRunge-Kutta method has to satisfy the stronger conditions in Table 6.1. To distinguish theseconditions from the ordinary conditions for IVP’s we refer to the conditions in Table 6.1 asoptimal control order conditions . A closer investigation yields that they are identical with theIVP conditions only up to order p = 2.

Table 6.1: Optimal control order conditions for Runge-Kutta methods up to order 4 (higherorder conditions require the validity of all conditions of lower order)

Order Conditions (ci =

aij, d j =

biaij)

p = 1

bi = 1

p = 2 di = 1

2

p = 3

cidi = 16 ,

bic2i = 13 , d2

i

bi= 1

3

p = 4

bic3i = 14 ,

biciaijc j = 18 ,

dic2i = 112 ,

diaijc j = 124 , cid2i

bi= 1

12 , d3i

b2i= 1

4 , biciaijdj

bj= 5

24 , diaijdj

bj= 1

8

Theorem 6.5.5 Let the assumptions (i)–(ii) in Assumption 6.5.4 be satisfied. Suppose bi > 0,i = 1, . . . , s holds for the coefficients of the Runge-Kutta method. Let the Runge-Kutta method be of optimal control order p, cf. Table 6.1.Then, for any sufficiently small step-size h > 0 there exists a strict local minimum of the

discretized optimal control problem.If dp−1u

dtp−1 is of bounded variation then

max0≤k≤N

xk − x(tk) + λk − λ(tk) + u∗(xk, λk) − u(tk) = O(h p).

If dp−1udtp−1 is Riemann-integrable then

max0≤k≤N

xk − x(tk) + λk − λ(tk) + u∗(xk, λk) − u(tk) = o(h p−1).

Herein, u∗(xk, λk) denotes a local minimum of the Hamilton function H(xk, u , λk) w.r.t. u.

Proof. The long and complicated proof can be found in Hager [Hag00].

Remark 6.5.6 The assumption bi > 0 is essential as Example 6.1.5 shows.

c 2006 by M. Gerdts

Page 193: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 193/243

Chapter 7

Selected Applications and Extensions

This chapter shows in an exemplified way several applications and extensions of the directdiscretization method of Chapter 6. Within each application it is necessary to solve dynamicoptimization problems.

7.1 Mixed-Integer Optimal Control

Many technical or economical processes lead to optimal control problems involving so-called‘continuous’ controls u ∈ U as well as ‘discrete’ controls v ∈ V . Herein, the control set U isassumed to be a closed convex subset of Rnu with nonempty interior whereas the control set

V is a discrete set with finitely many elements defined by

V = v1, . . . , vM vi ∈ Rnv , M ∈ N.

We consider

Problem 7.1.1 (Mixed-Integer Optimal Control Problem (MIOCP))

Minimize ϕ(x(t0), x(tf ))

w.r.t. x ∈ W 1,∞([t0, tf ],Rnx), u ∈ L∞([t0, tf ],R

nu), v ∈ L∞([t0, tf ],Rnv)

s.t. x(t) − f (t, x(t), u(t), v(t)) = 0 a.e. in [t0, tf ],

ψ(x(t0), x(tf )) = 0,

s(t, x(t)) ≤ 0 for all t ∈ [t0, tf ],

u(t) ∈ U for a.e. t ∈ [t0, tf ],

v(t) ∈ V for a.e. t ∈ [t0, tf ].

The functions ϕ : Rnx ×Rnx → R, f : [t0, tf ] ×Rnx ×Rnu × V → Rnx , ψ : Rnx ×Rnx → Rnψ , s :[t0, tf ]×Rnx → Rns are supposed to be sufficiently smooth w.r.t. to all arguments. Without lossof generality, it is assumed, that the initial time t0 and the final time tf are fixed. Furthermore,the restriction to ODEs is not essential. The following methods can be extended directly toDAEs.One way to solve the above optimal control problem is to formulate and solve necessary con-ditions provided by the well-known global minimum principle, cf. Ioffe and Tihomirov [IT79],Theorem 1, p. 234 and Girsanov [Gir72]. Unfortunately, for more complicated problems thisapproach usually turns out to be cumbersome and complex. Hence, we will concentrate onalternative approaches.

7.1.1 Branch&Bound

Another way, which was followed in Gerdts [Ger05e], is to apply the direct discretization ap-proach of Chapter 6 to MIOCP. This approach leads to a large scale finite dimensional nonlinearmixed-integer programming problem of type 3.10.1. This mixed-integer programming problem

185

Page 194: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 194/243

Page 195: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 195/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 187

By doing so, non-optimal values of the discrete control vGh,M within each main grid interval

[t(ti), t(ti+1)] will be ‘wiped out’. As in Ioffe and Tihomirov [IT79], p. 148 and in Lee etal. [LTRJ99] we define the variable time transformation t = t(τ ) by setting

t(τ ) := t0 + τ

t0 w(s)ds, τ ∈ [t0, tf ]

with tf

t0

w(s)ds = tf − t0,

i.e t(tf ) = tf . This transformation maps [t0, tf ] onto itself but changes the speed of runningthrough this interval. In particular, it holds

dt

dτ = w(τ ), τ ∈ [t0, tf ].

The function values w(τ ) with τ ∈ [τ i,j, τ i,j+1) are related to the length of the interval [t(τ i,j ), t(τ i,j+1)]according to τ i,j+1

τ i,j

w(τ )dτ = t(τ i,j+1) − t(τ i,j) in [τ i,j, τ i,j+1).

As a consequence the interval [t(τ i,j), t(τ i,j+1)] shrinks to the point t(τ i,j ) if

w(τ ) = 0 in [τ i,j, τ i,j+1).

Hence, w is restricted to the class of piecewise constant functions on the minor grid as depictedin Figure 7.2.

ti−1 ti ti+1 ti+2τ i−1,j τ i,j τ i+1,

w(τ )

Figure 7.2: Example of a function w controlling the length of every minor grid interval.

Motivated by the previous consideration concerning the interpretation of w the function w issubject to the following restrictions:

(i) w(τ ) ≥ 0 for all τ ;

(ii) w(τ ) is piecewise constant on the minor grid Gh,M ;

(iii) It holds

ti+1

ti

w(τ )dτ = ti+1 − ti = h, i = 0, . . . , N − 1.

c 2006 by M. Gerdts

Page 196: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 196/243

188 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

The first condition prohibits to run back in time and yields a non-decreasing function t(τ ). Thelatter condition is necessary to ensure that the parametrized time actually passes through theentire main grid interval [ti, ti+1]. The time points ti, i = 0, . . . , N serve as fixed time pointsin the original time t. Condition (iii) is essentially needed for mixed-integer optimal controlproblems, i.e. in the case that ‘continuous’ controls u and ‘discrete’ controls v are presentsimultaneously. The condition (iii) was not imposed in the papers [LTRJ97, LTRJ99, TJL99,Sib04, SR04] because only problems with a purely discrete control v and without additional‘continuous’ controls u were considered. In this case, condition (iii) is not needed. The reasonbehind it is that we need a sufficiently dense grid for the ‘continuous’ controls u in order toget a good approximation to MIOCP on Gh. If condition (iii) is not present then it will oftenhappen that after optimizing the discretized problem many main grid intervals will be shrunkento zero. In particular, for the numerical example from virtual test-driving at the end of thissection we will get solutions where the car is tunneling through the state constraints in themiddle of the test-course because all main grid points will be moved close to the start or theend of the test-course. Of course such a behavior is not desired and leads to wrong results.

Figure 7.3 illustrates the variable time transformation.

ti−1 ti ti+1τ i−1,j τ i,j τ i+1,j

w(τ )

t(τ ):

ti−1

ti

ti+1

ti+2

τ

Figure 7.3: Illustration of variable time transformation: Parametrization w (top) and corre-sponding time t = t(τ ) (bottom).

Notice, that the mapping τ →

t(τ ) is not invertible. But with the convention

t−1(s) := inf τ | s = t(τ ) (7.1.2)

it becomes invertible. The function vGh,M from (7.1.1) together with any w satisfying the

conditions (i)–(iii), e.g. w in Figure 7.2, correspond to a feasible discrete control v(t) ∈ V defined by

v(s) := vGh,M (t−1(s)), s ∈ [t0, tf ]

cf. Figure 7.4. Notice, that minor intervals with w(τ ) = 0 do not contribute to v(s).

c 2006 by M. Gerdts

Page 197: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 197/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 189

vGh,M (τ ): w(τ ):

Corresponding control v(s) = vGh,M (t−1(s)):

ti−1 ti ti+1t(τ i,1)= t(τ i,2) t(τ i+1,1)= t(τ i+1,2)

Figure 7.4: Back-transformation v (bottom) of variable time transformation for given w andfixed vGh,M

(top).

Vice versa, every piecewise constant discrete control v on the main grid Gh can be described byvGh,M

and some suitable w. Actually, the set of all functions that can be constructed by vGh,M

and w is larger than the set of all piecewise constant discrete-valued functions with values in V on the grid Gh. But it is smaller than the set of all such functions on the minor grid Gh,M . Forinstance, a function that switches from the value vM to v1 in some main grid interval can not berepresented. This is due to the preference of values given by the definition of the fixed function

vGh,M . If we would have defined vGh,M in each main grid interval starting with the value vM andending with v1 we would get a different set of representable functions. However, for h → 0 anydiscrete-valued function with finitely many jumps in (t0, tf ) can be approximated arbitrarilyclose (in the L1-norm) by vGh,M

and a suitable w. Thus, the suggested transformation forsufficiently small h will yield good approximations for mixed-integer optimal control problemswith the optimal v having only finitely many jumps.

Since our intention is to create an optimal v we introduce w as a new control subject to

w ∈ W :=

w ∈ L∞([t0, tf ],R

nv)

w(τ ) ≥ 0,w piecewise constant on Gh,M ,

ti+1

ti w(τ )dτ = ti+1 − ti ∀i

.

The transformed mixed-integer optimal control problem (TMIOCP) reads as

Minimize ϕ(x(t0), x(tf ))w.r.t. x ∈ W 1,∞([t0, tf ],R

nx), u ∈ L∞([t0, tf ],Rnu), w ∈ L∞([t0, tf ],R

nv)s.t. x(τ ) − w(τ )f (τ, x(τ ), u(τ ), vGh,M

(τ )) = 0 a.e. in [t0, tf ],

s(τ, x(τ )) ≤ 0 in [t0, tf ],ψ(x(t0), x(tf )) = 0

u(τ ) ∈ U a.e. in [t0, tf ]w

∈ W c 2006 by M. Gerdts

Page 198: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 198/243

Page 199: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 199/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 191

Problem 7.1.3

Minimize ϕ(x0,0, xN,0)w.r.t. xi,j, ui, wi,j ,s.t. equations (7.1.3)

−(7.1.5),

s(ti, xi,0) ≤ 0 i = 0, . . . , N ,ψ(x0,0, xN,0) = 0,

ui ∈ U i = 0, . . . , N − 1,wi,j ≥ 0, i = 0, . . . , N − 1, j = 0, . . . , M − 1,

M −1 j=0

wi,j = M, i = 0, . . . , N − 1.

This problem can be solved numerically by a SQP method.

Remark 7.1.4 The state constraint s(t, x(t)) ≤ 0 is only evaluated on the main grid and not on the minor grid. This keeps the number of constraints small. Likewise, uh is kept piecewise

constant on the main grid and not on the minor grid. This keeps the number of variables small. This approach should be sufficient in view of a possible convergence result as h → 0.Nevertheless, it is straightforward to construct an alternative method discretizing everything on the minor grid.

7.1.3 Numerical Results

We will consider a particular optimal control problem arising from automobile test-driving withgear shifts. Herein, one component of the control is discrete.

7.1.3.1 Car Model

lr

lf

eSP

F sr

F lr

vr

αr

v

αψ

F lf

F sf

αf

vf

δ

F Ax

F Ay

(x, y)

Figure 7.5: Geometrical description of the single-track car model.

c 2006 by M. Gerdts

Page 200: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 200/243

192 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

We use the single-track car model to be described below in detail. It is a simplified car model,which is commonly used in automobile industry for basic investigations of the dynamical behaviorof cars, cf., e.g., Mayr [May91], Risse [Ris91], Neculau [Nec92], Moder [Mod94].

The simplifying assumption made to derive the single-track car model is that the rolling and

pitching behavior of the car body can be neglected, that is, the roll angle and the pitch angle aresmall. This allows to replace the two wheels on the front and rear axis by a virtual wheel locatedin the center of the respective axis. Furthermore, due to the simplifying assumptions it can beassumed that the car’s center of gravity is located on the roadway and therefore, it is sufficient toconsider the motion of the car solely in the horizontal plane. The upcoming car model includesfour control variables for the driver: the steering angle velocity |wδ| ≤ 0.5 [rad/s], the totalbraking force 0 ≤ F B ≤ 15000 [N ], the gear µ ∈ 1, 2, 3, 4, 5, and the accelerator pedal positionφ ∈ [0, 1]. The configuration of the car is depicted in Figure 7.5.

Herein, (x, y) denotes the center of gravity in a reference coordinate system, v, vf , vr denote thevelocity of the car and the velocities of the front and rear wheel, respectively, δ,β,ψ denotethe steering angle, the side slip angle, and the yaw angle, respectively, αf , αr denote the slip

angles at front and rear wheel, respectively, F sf , F sr denote the lateral tire forces (side forces) atfront and rear wheel, respectively, F lf , F lr denote the longitudinal tire forces at front and rearwheel, respectively, lf , lr, eSP denote the distances from the center of gravity to the front andrear wheel, and to the drag mount point, respectively, F Ax, F Ay denote the drag according toair resistance and side wind, respectively, and m denotes the mass of the car.

The equations of motion are given by the following system of ordinary differential equations

x = v cos(ψ − β ), (7.1.6)

y = v sin(ψ − β ), (7.1.7)

v =

1

m (F lr − F Ax)cos β + F lf cos(δ + β ) − (F sr − F Ay)sin β

−F sf sin(δ + β )

, (7.1.8)

β = wz − 1

m · v

(F lr − F Ax)sin β + F lf sin(δ + β )

+ (F sr − F Ay)cos β + F sf cos(δ + β )

, (7.1.9)

ψ = wz, (7.1.10)

wz = 1

I zz

F sf · lf · cos δ − F sr · lr − F Ay · eSP + F lf · lf · sin δ

, (7.1.11)

δ = wδ. (7.1.12)

The lateral tire forces are functions of the respective slip angles (and the tire loads, which areconstant in our model). A famous model for the lateral tire forces is the ‘magic formula’ of Pacejka [PB93]:

F sf (αf ) = Df sin(C f arctan (Bf αf − E f (Bf αf − arctan(Bf αf )))) ,

F sr(αr) = Dr sin (C r arctan (Brαr − E r (Brαr − arctan(Brαr)))) ,

compare Figure 7.6. Herein, Bf , Br, C f , C r, Df , Dr, E f , E r are constants depending on the tire.

c 2006 by M. Gerdts

Page 201: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 201/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 193

-4000

-2000

0

2000

4000

-0.4 -0.3 -0.2 -0.1 0 0.1 0.2 0.3 0.4

-4000

-2000

0

2000

4000

-0.4 -0.3 -0.2 -0.1 0 0.1 0.2 0.3 0.4

αf [rad] αr [rad]

F s f

( α f

)

[

N

]

F s r

( α r

)

[ N

]

Figure 7.6: Lateral tire forces at front (left) and rear (right) wheel as functions of the slip anglecomputed by the ‘magic formula’ of Pacejka [PB93].

According to Mitschke [Mit90], p. 23, (notice the opposite sign in the definition of β ) the slipangles are given by

αf = δ − arctan

lf

ψ − v sin β

v cos β

, αr = arctan

lr

ψ + v sin β

v cos β

.

The drag due to air resistance is modeled by

F Ax = 1

2 · cw · ρ · A · v2,

where cw is the air drag coefficient, ρ is the air density, and A is the effective flow surface. Inthis article, it is assumed that there is no side wind, i.e. F Ay = 0.

In the sequel, we assume that the car has rear wheel drive. Then, the longitudinal tire force atthe front wheel is given by

F lf = −F Bf − F Rf ,

where F Bf is the braking force at the front wheel and F Rf denotes the rolling resistance at thefront wheel. The longitudinal tire force at the rear wheel is given by

F lr = M wheel(φ, µ)

R − F Br − F Rr

where M wheel(φ, µ) is the torque resulting from the drive-train applied to the rear wheel. Again,F Br is the braking force at the rear wheel and F Rr denotes the rolling resistance at the rearwheel. According to Neculau [Nec92] it holds

M wheel(φ, µ) = ig(µ) · it · M mot(φ, µ), (7.1.13)

where

M mot(φ, µ) = f 1(φ) · f 2(wmot(µ)) + (1 − f 1(φ))f 3(wmot(µ))

denotes the motor torque and

wmot(µ) = v · ig(µ) · it

R

(7.1.14)

c 2006 by M. Gerdts

Page 202: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 202/243

194 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

denotes the rotary frequency of the motor depending on the gear µ. Notice, that this relation isbased on the assumption that the longitudinal slip can be neglected. The functions f 1, f 2 andf 3 are given by

f 1(φ) = 1 − exp(−3φ),

f 2(wmot) = −37.8 + 1.54 · wmot − 0.0019 · w2mot,f 3(wmot) = −34.9 − 0.04775 · wmot.

The total braking force F B controlled by the driver is distributed on the front and rear wheelssuch that F Bf + F Br = F B holds. Here, we chose

F Bf = 2

3F B, F Br =

1

3F B .

Finally, the rolling resistance forces are given by

F Rf = f R(v) · F zf , F Rr = f R(v) · F zr ,

where

f R(v) = f R0 + f R1

v

100 + f R4 v

1004

(v in [km/h]),is the friction coefficient and

F zf = m · lr · g

lf + lr, F zr =

m · lf · g

lf + lr

denote the static tire loads at the front and rear wheel, respectively, cf. Risse [Ris91].

For the upcoming numerical computations we used the parameters summarized in Table 7.1.

Table 7.1: Parameters for the single-track car model (partly taken from Risse [Ris91]).

Parameter Value Descriptionm 1239 [kg] mass of carg 9.81 [m/s2] acceleration due to gravitylf /lr 1.19016/1.37484 [m] dist. center of gravity to front/rear wheeleSP 0.5 [m] dist. center of gravity to drag mount pointR 0.302 [m] wheel radiusI zz 1752 [kgm2] moment of inertiacw 0.3 air drag coefficientρ 1.249512 [N/m2] air densityA 1.4378946874 [m2] effective flow surfaceig(1) 3.91 first gear

ig(2) 2.002 second gearig(3) 1.33 third gearig(4) 1.0 fourth gearig(5) 0.805 fifth gearit 3.91 motor torque transmissionBf /Br 10.96/12.67 Pacejka-model (stiffness factor)C f = C r 1.3 Pacejka-model (shape factor)Df /Dr 4560.40/3947.81 Pacejka-model (peak value)E f = E r −0.5 Pacejka-model (curvature factor)f R0/f R1/f R4 0.009/0.002/0.0003 coefficients

c 2006 by M. Gerdts

Page 203: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 203/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 195

7.1.3.2 Test-Course

The discussion is restricted to the double-lane-change manoeuvre being a standardized manoeu-vre in automobile industry. The driver has to manage an offset of 3.5 [m] and afterwards he hasto reach the original track, see Figure 7.7. The double-lane-change manoeuvre can be understood

as a model for a jink caused by a suddenly occurring obstacle on the road.

Figure 7.7: Measurements of the track and boundaries P l and P u (dashed), cf. Zomotor [Zom91]

The boundaries of the test-course will play the role of state constraints in a suitable optimalcontrol problem later on and are described by continuously differentiable functions P l(x) (lowerboundary) and P u(x) (upper boundary), which are piecewise defined cubic polynomials. For acar with width B = 1.5 [m] we find, cf. Moder [Mod94], Gerdts [Ger01a]:

P l(x) =

0, if x ≤ 44 ,4 · h2 · (x − 44)3, if 44 < x ≤ 44.5 ,4 · h2 · (x − 45)3 + h2, if 44.5 < x ≤ 45 ,h2, if 45 < x ≤ 70 ,4 · h2 · (70 − x)3 + h2, if 70 < x ≤ 70.5 ,4 · h2 · (71 − x)3, if 70.5 < x ≤ 71 ,0, if x > 71 ,

, (7.1.15)

P u(x) =

h1, if x ≤ 15 ,4 · (h3 − h1) · (x − 15)3 + h1, if 15 < x ≤ 15.5 ,4 · (h3 − h1) · (x − 16)3 + h3, if 15.5 < x ≤ 16 ,h3, if 16 < x

≤ 94 ,

4 · (h3 − h4) · (94 − x)3 + h3, if 94 < x ≤ 94.5 ,4 · (h3 − h4) · (95 − x)3 + h4, if 94.5 < x ≤ 95 ,h4, if x > 95 ,

(7.1.16)

where h1 = 1.1 · B + 0.25, h2 = 3.5, h3 = 1.2 · B + 3.75, h4 = 1.3 · B + 0.25.

7.1.3.3 Driver Model

The driver is modeled by formulation of a suitable state constrained optimal control problem withfree final time. The required observation of the marking cones results in two state constraints

y(t)≤

P u(x(t))−

B/2, y(t)≥

P l(x(t)) + B/2, (7.1.17)

c 2006 by M. Gerdts

Page 204: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 204/243

196 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

where (y(t), x(t)) denotes the position of the car’s center of gravity at time t and B the car’swidth. Let the initial state of the car at time t0 = 0 on the track be prescribed with exceptionof the initial y-position:

(x(0), y(0), v(0), β (0), ψ(0), wz (0), δ (0)) = (

−30,free, 10, 0, 0, 0, 0). (7.1.18)

To ensure, that the car completes the course, we impose additional boundary conditions at thefinal time tf :

x(tf ) = 140, ψ(tf ) = 0. (7.1.19)

The latter condition ensures, that the longitudinal axis of the car is parallel to the boundaryof the test-course at tf . The intention of this condition is to ensure that the driver couldcontinue his drive at least for a short time period after tf has been reached without violatingthe boundaries of the course too severely.Finally, the driver is modeled by minimizing a linear combination of the steering effort and thefinal time. Roughly speaking, this is a compromise between driving fast (minimizing time) and

driving safely (minimizing steering effort).Summarizing, the double-lane-change manoeuvre is modeled by the following optimal controlproblem:

Minimize tf +

tf

0wδ(t)2dt

s.t. differential equations (7.1.6)-(7.1.12) a.e. in [0, tf ],boundary conditions (7.1.18) and (7.1.19),state constraints (7.1.17) for all t ∈ [0, tf ],control constraints wδ(t) ∈ [−0.5, 0.5], F B (t) ∈ [0, 15000],φ ∈ [0, 1], µ(t) ∈ 1, 2, 3, 4, 5 for all t ∈ [0, tf ].

Remark 7.1.5 For the applicability of the Branch&Bound method it is necessary to relax the discrete function ig(µ). This can be done by constructing an interpolating natural cubic spline as it was done in Gerdts [Ger05e].

7.1.4 Results

We compare the Branch&Bound method and the variable time transformation method. For thenumerical calculations we used the software package SODAS, cf. Gerdts [Ger01a, Ger03a], withthe classical Runge-Kutta discretization scheme of order four with piecewise constant controlapproximations on an equidistant grid. All computations concerning the Branch&Bound methodwere performed on a Pentium 3 processor with 750 MHz. All computations concerning thevariable time transformation method were performed on a Pentium mobile processor with 1.6GHz processing speed.In all cases the optimal braking forces and the optimal gas pedal positions compute to F B ≡ 0and ϕ ≡ 1, respectively.

Table 7.2 summarizes the numerical results of the Branch&Bound method for N = 20 gridpoints. Herein, the Branch&Bound algorithm is stopped after #NLP nodes of the search treeare generated. Recall, that each node corresponds to solving one relaxed discretized optimalcontrol problem. At termination the lower and upper bound (columns LOWER and UPPER) arecomputed in order to obtain an estimate for the exact optimal objective function value. Withthese bounds the accuracy of the solution obtained at termination is given by column GAP,where GAP=(UPPER-LOWER)/LOWER. This procedure allows to investigate how good thecurrently best solutions are after a fixed number of iterations. Without artificial stopping, for

c 2006 by M. Gerdts

Page 205: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 205/243

Page 206: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 206/243

198 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

Figure 7.8 shows the numerical solution for the variable time transformation method for N = 20.The variable time transformation method only needs 2 minutes and 1.154 seconds to solve theproblem with objective function value 6.774669 and final time tf = 6.772516 [s]. Recall, thatthe Branch&Bound method needed 23 minutes and 52 seconds! The lower objective functionvalue for the variable time transformation method is due to additional degrees of freedom inProblem 7.1.3, since the method allows to switch the discrete control even within some maingrid interval. The switching points of the discrete control in the Branch&Bound method wererestricted to the main grid points. Nevertheless, the qualitative switching structure of theoptimal discrete control µ is identical for both approaches.

-0.03

-0.02

-0.01

0

0.01

0.02

0.03

0 0.2 0.4 0.6 0.8 1

1

2

3

4

5

0 0.2 0.4 0.6 0.8 1

-2

-1

0

1

2

3

4

5

6

7

-20 0 20 40 60 80 100 120 140

y ( t )

[ m ]

x(t) [m]

w δ

( t )

µ ( t )

normalized t [s] normalized t [s]

Figure 7.8: Numerical result for N = 20: Path (x(t), y(t)) of the car’s center of gravity (top),steering angle velocity wδ(t) (bottom, left), and gear shift µ(t) (bottom, right).

Figure 7.9 shows the numerical solution for the variable time transformation method for N = 40.The Branch&Bound method needed 232 hours, 25 minutes and 31 seconds of CPU time tosolve the problem and yields the optimal objective function value 6.791374 and final time t f =6.786781 [s]. The variable time transformation method only needs 9 minutes and 39.664 secondsto solve the problem with objective function value 6.787982 and final time t f = 6.783380 [s].

c 2006 by M. Gerdts

Page 207: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 207/243

7.1. MIXED-INTEGER OPTIMAL CONTROL 199

-0.05

-0.04

-0.03

-0.02

-0.01

0

0.01

0.02

0.03

0.04

0.05

0.06

0 0.2 0.4 0.6 0.8 1

1

2

3

4

5

0 0.2 0.4 0.6 0.8 1

-2

-1

0

1

2

3

4

5

6

7

-20 0 20 40 60 80 100 120 140

y ( t )

[ m ]

x(t) [m]

w δ

( t )

µ ( t )

normalized t [s] normalized t [s]

Figure 7.9: Numerical result for N = 40: Path (x(t), y(t)) of the car’s center of gravity (top),steering angle velocity wδ(t) (bottom, left), and gear shift µ(t) (bottom, right).

Figure 7.10 shows the numerical solution for N = 80. The Branch&Bound method did notterminate in a reasonable time. The variable time transformation method only needs 65 minutesand 3.496 seconds to solve the problem with objective function value 6.795366 and final timetf = 6.789325 [s].

It has to be mentioned that the calculation of gradients in the SQP method involved in thevariable time transformation method for simplicity was done by finite differences. Using a moresophisticated approach as in Chapter 6 would even lead to a substantial reduction of CPUtime for the variable time transformation method. These computational results show that thevariable time transformation method is particularly well-suited for mixed-integer optimal controlproblems and is much less expensive than the Branch&Bound method.

c 2006 by M. Gerdts

Page 208: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 208/243

200 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

-0.06

-0.04

-0.02

0

0.02

0.04

0.06

0.08

0 0.2 0.4 0.6 0.8 1

1

2

3

4

5

0 0.2 0.4 0.6 0.8 1

-2

-1

0

1

2

3

4

5

6

7

-20 0 20 40 60 80 100 120 140

y ( t )

[ m ]

x(t) [m]

w δ

( t )

µ ( t )

normalized t [s] normalized t [s]

Figure 7.10: Numerical result for N = 80: Path (x(t), y(t)) of the car’s center of gravity (top),steering angle velocity wδ(t) (bottom, left), and gear shift µ(t) (bottom, right).

c 2006 by M. Gerdts

Page 209: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 209/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 201

7.2 Open-Loop-Real-Time Control

In practical applications the optimal control problem depends on system parameters. For ex-ample, for the computation of optimal flight paths of a space shuttle mission a common modelfor the air density depends on the altitude and certain parameters describing a standard atmo-

sphere. Of course, in reality the air density differs from the model and the actual air density canbe viewed as a perturbation of the model. This in turn implies that the computed optimal flightpath for the standard atmosphere is not optimal anymore for the actual air density. Accordingly,the flight path has to be adapted to the perturbed situation in an optimal way. One way toachieve this would be to re-compute the optimal flight path for the perturbed model. However,for time critical processes the re-computation of an optimal solution for the perturbed problem,e.g. by the direct discretization method, may be not fast enough to provide an optimal solutionin real-time. Therefore, an alternative method is needed which is capable of providing at leastan approximation of the optimal solution for the perturbed problem in real-time. In the sequel amethod based on the parametric sensitivity analysis of the underlying optimal control problemis suggested to calculate such real-time approximations.

We consider

Problem 7.2.1 (Perturbed DAE Optimal Control Problem OC P ( p))

Minimize ϕ(x(t0), x(tf ), p)

s.t. F (t, x(t), x(t), u(t), p) = 0nx a.e. in [t0, tf ],

ψ(x(t0), x(tf ), p) = 0nψ,

c(t, x(t), u(t), p) ≤ 0nc a.e. in [t0, tf ],

s(t, x(t), p) ≤ 0ns in [t0, tf ].

Notice, that p ∈ Rnp is not an optimization variable. Let ˆ p ∈ Rnp denote a fixed nominal parameter . The corresponding optimal control problem OCP (ˆ p) is called nominal problem .

For a given parameter p we may apply the reduced discretization approach to solve the problemnumerically.

Problem 7.2.2 (Reduced Discretization DOCP ( p))Find z = (x0, w) ∈Rnx+M such that the objective function

ϕ(X 0(z, p), X N (z, p), p)

is minimized subject to

ψ(X 0(z, p), X N (z, p), p) = 0nψ,

c(t j , X j (z, p), uM (t j ; w), p) ≤ 0nc , j = 0, 1, . . . , N ,s(t j, X j(z, p), p) ≤ 0ns , j = 0, 1, . . . , N .

Herein, X 0 is a function that provides a consistent initial value for x0, control parametrizationw, and parameter p. The values X j, j = 1, . . . , N are given by a suitable integration scheme,e.g. a one-step method:

X j+1(z, p) = X j(z, p) + h j Φ(t j, X j (z, p),w,p,h j ), j = 0, 1, . . . , N −

1.

c 2006 by M. Gerdts

Page 210: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 210/243

202 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

According to (6.1.2) the discretized control is given by

uM (·) =M

i=1

wiBi(·)

with basis functions Bi, e.g. B-Splines.

Suppose that DOCP ( p) is solvable for any p ∈ Rn, let z( p) = (x0( p), w( p)) denote the opti-mal solution and z(ˆ p) the nominal solution . In particular, the optimal discretized control forDOCP ( p) is given by

uM (·) =M

i=1

wi( p)Bi(·).

The underlying assumption to obtain a real-time update formula, i.e. a formula or methodthat is capable of providing an approximation for the optimal solution z( p) in real-time, isthat there exists a neighborhood ∅ = N (ˆ p) ⊆ Rnp of the nominal parameter ˆ p, such that

the perturbed problem DOCP ( p) possesses an optimal solution z( p) for all p ∈ N (ˆ p), whichis continuously differentiable w.r.t. p in N (ˆ p). Sufficient conditions for differentiability arestated in Theorem 3.7.3. In case of differentiability, Taylor expansion yields the real-time update

formula

z( p) ≈ z(ˆ p) + dz(ˆ p)

dp · ( p − ˆ p), (7.2.1)

cf. equation (3.7.4). Application of this formula to the discretized control yields a real-timeapproximation of the optimal perturbed control according to

uM (·) ≈M

i=1 wi(ˆ p) + dwi(ˆ p)

dp ( p − ˆ p)Bi(·). (7.2.2)

Please notice, that only matrix-vector multiplications and vector-vector summations are neededto evaluate the real-time update formulae. In the case of an one-dimensional parameter, theseoperations can be performed within micro seconds or even below on standard computers. Hence,the real-time update formulae provide an extremely fast approximation of the optimal solutionof the perturbed optimal control problem. It remains to compute the sensitivity differentialdz(ˆ p)/dp. Two approaches for the calculation of the sensitivity differential are described in thesequel: the brute-force method and the approach via sensitivity analysis. In either case, thecalculation of the sensitivity differential takes place offline and may be very time consuming.Nevertheless, if the sensitivity differential is obtained, the on-line evaluation of the real-timeupdate formulas (7.2.1) and (7.2.2), respectively, is extremely cheap as mentioned above.

Remark 7.2.3 Conditions for the solution differentiability of the infinite dimensional optimal control problem subject to ODEs in appropriate Banach spaces can be found in Maurer and Pesch [MP94a, MP94b, MP95a], Malanowski and Maurer [MM96, MM98, MM01], and Maurer and Augustin [AM01, MA01]. Numerical approaches based on a linearization of the necessary op-timality conditions of the optimal control problem are discussed in Pesch [Pes78, Pes79, Pes89a,Pes89b].

7.2.1 Brute-force method

The idea of the brute-force method is very simple. Let pi be the i−th component of the parametervector p = ( p1, . . . , pnp) and ei

∈Rnp the i

−th unit vector. The sensitivity differential dz(ˆ p)/dpi

c 2006 by M. Gerdts

Page 211: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 211/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 203

is approximated by, e.g., the central finite difference scheme

dz(ˆ p)

dpi≈ z(ˆ p + hei) − z(ˆ p − hei)

2h , (7.2.3)

where z (ˆ p ± hei) denotes the optimal solution of DOCP (ˆ p ± hei). Thus, the calculation of thesensitivity differential dz(ˆ p)/dp requires the solution of 2n p discretized optimal control problems.Depending on n p this is a potentially large number of nonlinear programming problems to besolved. In addition, the choice of the step size h is crucial. Despite this disadvantages, one hasthe advantage, that a very good initial estimate for the SQP method is known: the nominalsolution z(ˆ p). In addition, this approach may be preferable to the subsequent method, if theoptimal solution can not be calculated very accurately, which is often the case for large scaleproblems.

7.2.2 Real-time approximation via sensitivity analysis

Buskens and Maurer [BM96, BM01b, BM01a] established a method for ODE optimal control

problems, which allows to calculate an approximation of the optimal solution of a perturbedoptimal control problem in real-time. This method is based on a sensitivity analysis of thediscretized optimal control problem w.r.t. to the perturbation parameters p. Buskens andGerdts [GB01] extended this method to optimal control problems subject to semi explicit index-1 DAEs.The discretized optimal control problem DOCP (ˆ p) is a parametric nonlinear program as inProblem 3.7.1 with objective function and constraints according to (6.1.12)-(6.1.14).

Suppose that the SQP method applied to DOCP (ˆ p) converges to a local minimum z = z(ˆ p)with Lagrange multiplier µ and that the assumptions of Theorem 3.7.3 hold at z. Then, Theo-rem 3.7.3 guarantees the differentiability of the mapping p → z( p) in some neighborhood of ˆ p.Furtheremore, a solution of the linear equation (3.7.1) yields the sensitivity differential dz

dp (ˆ p) at

the solution z.The solution of (3.7.1) requires the second derivatives L

zz and Lzp of the Lagrange function,

the derivatives H z, Gz in (6.1.15)-(6.1.16), and the corresponding derivatives H p, G

p at z. Moreprecisely, only the derivatives of the active constraints are neeeded. Due to high computationalcosts, most implementations of SQP methods, e.g. NPSOL of Gill et. al. [GMSW98], do not usethe Hessian of the Lagrangian within the QP subproblem. Instead, the Hessian is replaced by amatrix computed with the modified BFGS update formula (3.8.7). It is well known, that even inthe optimal solution the matrix computed by the BFGS update formula may differ substantiallyfrom the exact Hessian and therefore can not be used for sensitivity computations. Boggset. al. [BTW82] stated a necessary and sufficient condition for Q-superlinear convergence of quasi-Newton methods for equality constrained nonlinear optimization problems: Under certain

assumptions, the sequence z(k)

converges Q-superlinearly to z, if and only if

P (z(k), ˆ p) (Bk − Lzz (z, µ, ˆ p)) d(k)

d(k) → 0,

where Bk is the BFGS update matrix and P is a projector on the tangent space of the constraints.This means, that the BFGS update is only a good approximation of the Hessian when projectedon the linearized constraints at iterate z(k) in direction d(k).Consequently, the Hessian L

zz as well as the quantities Lzp, H z, G

z , H p, and G p have to be

reevaluated after the SQP method converged.The computation of the derivatives H z, G

z, H p, and G p can be done efficiently by either the

sensitivity equation approach of Section 6.2.1 or the adjoint equation approach of Sections 6.2.2

c 2006 by M. Gerdts

Page 212: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 212/243

204 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

and 6.2.3. Notice that these approaches usually are less expensive and more reliable than finitedifference approximations.

One idea to approximate the derivatives Lzz and L

zp is to use the finite difference approximations

Lzizj(z, µ, ˆ p) ≈ 1

4h2 (L(z + hei + he j, µ, ˆ p) − L(z − hei + he j , µ, ˆ p)

−L(z + hei − he j , µ, ˆ p) + L(z − hei − he j, µ, ˆ p)) ,

Lzi pj

(z, µ, ˆ p) ≈ 1

4h2 (L(z + hei, µ, ˆ p + he j ) − L(z − hei, µ, ˆ p + he j)

−L(z + hei, µ, ˆ p − he j ) + L(z − hei, µ, ˆ p − he j ))

(7.2.4)

with appropriate unity vectors e i, e j . For each evaluation of the Lagrange function L one DAE

has to be solved numerically. Hence, this approach leads to a total number of 4 nz(nz+1)2 + 4nzn p

DAE evaluations if the symmetry of the Hessian Lzz is exploited.

An alternative approach is to exploit the information obtained by the sensitivity equation. Recall

that one evaluation of the DAE and the corresponding sensitivity equation for the NLP variablesz provides the gradient of the objective function and the Jacobian of the constraints and withthis information the gradient of the Lagrangian is obtained easily. Again, a finite differenceapproximation with these gradients yields the second order derivatives according to

Lzzi

(z, µ, ˆ p) ≈ Lz(z + hei, µ, ˆ p) − L

z(z − hei, µ, ˆ p)

2h ,

Lzpj

(z, µ, ˆ p) ≈ Lz(z, µ, ˆ p + he j ) − L

z(z, µ, ˆ p − he j )

2h .

This time, only 2nz + 2n

p DAEs including the corresponding sensitivity equations (6.2.3) w.r.t.

z have to be solved. Since the evaluation of nz nonlinear DAEs in combination with the corre-sponding linear sensitivity DAE with nz columns is usually much cheaper than the evaluation of nz

2 nonlinear DAEs, the second approach is found to be much faster than a pure finite differenceapproximation (7.2.4).

The main drawback of this method based on a sensitivity analysis of (DOCP ( p) is, that avery accurate solution of DOCP (ˆ p) is needed in view of getting good results for the sensitivitydifferentials. As a rule of thumb, the feasibility and optimality tolerance of the SQP method,compare Gill et al. [GMSW98], should be smaller than 10−9, if the derivatives are computed byfinite differences.

7.2.3 Numerical Example: Emergency LandingDuring the ascent phase of a winged two-stage hypersonic flight system some malfunction ne-cessitates to abort the ascent shortly after separation. The upper stage of the flight system isstill able to manoeuvre although the propulsion system is damaged, cf. Mayrhofer and Sachs[MS96], Buskens and Gerdts [BG00, BG03]. For security reasons an emergency landing trajec-tory with maximum range has to be found. This leads to the following optimal control problemfor t ∈ [0, tf ]:

Minimize

Φ(C L, µ , tf ) =

−Λ(tf ) − Λ(0)

Λ(0) 2

− Θ(tf ) − Θ(0)

Θ(0) 2

(7.2.5)

c 2006 by M. Gerdts

Page 213: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 213/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 205

subject to the ODE for the velocity v, the inclination γ , the azimuth angle χ, the altitude h,the latitude Λ, and the longitude Θ

v = −D(v, h; C L) 1

m − g(h)sin γ +

+ ω2 cos Λ(sin γ cosΛ − cos γ sin χ sin Λ)R(h), (7.2.6)

γ = L(v, h; C L)cos µ

mv −

g(h)

v − v

R(h)

cos γ +

+ 2ω cos χ cos Λ + ω2 cos Λ(sin γ sin χ sin Λ + cos γ cos Λ)R(h)

v , (7.2.7)

χ = L(v, h; C L) sin µ

mv cos γ − cos γ cos χ tanΛ

v

R(h) +

+ 2ω(sin χ cosΛtan γ − sinΛ) − ω2 cosΛsinΛcos χ R(h)

v cos γ , (7.2.8)

h = v sin γ, (7.2.9)

Λ = cos γ sin χ v

R(h), (7.2.10)

Θ = cos γ cos χ v

R(h)cosΛ, (7.2.11)

with functions

L(v,h,C L) = q (v, h) F C L, ρ(h) = ρ0 exp(−βh),D(v,h,C L) = q (v, h) F C D(C L), R(h) = r0 + h,

C D(C L) = C D0 + k C L2, g(h) = g0(r0/R(h))2,

q (v, h) = 12ρ(h)v

2

(7.2.12)

and constants

F = 305, r0 = 6.371 · 106, C D0 = 0.017,k = 2, ρ0 = 1.249512 · (1 + p), β = 1/6900,

g0 = 9.80665, ω = 7.27 · 10−5, m = 115000.(7.2.13)

The parameter p will be used to model perturbations in the air density ρ(h) in (7.2.12). For theunperturbed problem it holds p = ˆ p = 0.Since the propulsion system is damaged, the mass m remains constant. Box constraints for thetwo control functions lift coefficient C L and angle of bank µ are given by

0.01 ≤ C L ≤ 0.18326,

−π

2 ≤ µ ≤ π

2.

(7.2.14)

The initial values for the state correspond to a starting position above Bayreuth/Germany

v(0)γ (0)χ(0)h(0)Λ(0)Θ(0)

=

2150.54529000.15201817702.268927988933900.0000000.86515971020.1980948701

. (7.2.15)

c 2006 by M. Gerdts

Page 214: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 214/243

206 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

An additional restriction is given by the terminal condition

h(tf ) = 500 − ε. (7.2.16)

The parameter ε will be used to model perturbations in the final altitude h(tf ). For the unper-turbed problem it holds ε = ε0 = 0.

The final time tf is assumed to be free and thus tf is an additional optimization variable. Finally,the dynamic pressure constraint

q (v, h) ≤ q max (7.2.17)

has to be obeyed. Here, we used the value q max = 60000 [N/m2].

The infinite dimensional optimal control problem is discretized by the reduced discretizationapproach. We used the classical fourth order Runge-Kutta scheme for time integration. Thecontrol is approximated by a continuous and piecewise linear function.

Figure 7.11 shows the numerical solution for 151 discretization points of the unperturbed problem

with p = ˆ p = 0 and ε = ε0 = 0. The dynamic pressure constraint is active in the normalizedtime interval [0.18, 0.186666666].

Table 7.4: Results for the emergency landing manoeuvre with and without dynamic pressureconstraint for different SQP methods.

method dyn. tf objective CPU nom. CPU total eigenvalue evalpress. in [s] in [s] in [s] red. Hessian

NPSOL no 728.8493380 −0.7758097 60.66 93.40 3.97 · 10−5 500NPSOL yes 727.1079063 −0.7649252 60.04 94.54 4.31 · 10−5 382

Table 7.4 summarizes data for the emergency landing manoeuvre with and without dynamicpressure constraint. The columns ‘tf ’ and ‘objective’ denote the final time and the objectivefunction value, respectively, of the nominal solution of the discretized optimal control problems.The computations were performed on a personal computer with Pentium 4 processor and 2.66GHz. The column ‘CPU nom.’ indicates the CPU time needed for the nominal problem, whereas‘CPU total’ includes in addition the sensitivity analysis of the nonlinear programming problem.The column ‘eigenvalue red. Hessian’ denotes the smallest eigenvalue of the reduced Hessian. If this entry is positive then the second order sufficient conditions in Theorem 3.6.8 are satisfiedfor the nonlinear optimization problem. In particular the Hessian of the Lagrangian is positivedefinite on the kernel of the linearized active constraints. The column ‘eval’ denotes the number

of objective function and constraint evaluations of the employed SQP method. The optimalitytolerance and feasibility tolerance for Npsol were both set to 10−12.

Table 7.5 summarizes data for the emergency landing manoeuvre with and without dynamicpressure constraint.

Table 7.5: Sensitivities of the nominal final time tf w.r.t. perturbation parameters p and ε forthe emergency landing manoeuvre with and without dynamic pressure constraint.

method dyn. press. dtf /dp dtf /dεNPSOL no 91.8538973 0.0145585106NPSOL yes 92.3851543 0.0141449020

c 2006 by M. Gerdts

Page 215: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 215/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 207

no dynamic pressure con.with dynamic pressure con.

0.860.880.90.92

0.94

0.040.08

0.120.16

0.2

0

10000

20000

30000

40000

50000

0.06

0.08

0.1

0.12

0.14

0.16

0.18

0.2

0 0.2 0.4 0.6 0.8 1

l i f t c o e f f i c i e n t

normalized time

no dynamic pressure con.with dynamic pressure con.

-0.2

0

0.2

0.4

0.6

0.8

1

1.2

0 0.2 0.4 0.6 0.8 1

a n g l e

o f b a n k

[ r a d

]

normalized time

no dynamic pressure con.with dynamic pressure con.

Λ [rad] Θ [rad]

h [m]

Figure 7.11: Comparison of the numerical nominal solutions for the problem with and withoutdynamic pressure constraint: 3D plot of the flight path (top) and the approximate optimalcontrols lift coefficient C L and angle of bank µ (bottom, normalized time scale) for 151 gridpoints.

Figures 7.12-7.15 show the sensitivities of the nominal controls C L and µ w.r.t. the perturbationparameters p and ε. For the problem with dynamic pressure constraint the sensitivities jump atpoints where the state constraint becomes active resp. inactive.

c 2006 by M. Gerdts

Page 216: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 216/243

208 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

-0.14

-0.12

-0.1

-0.08

-0.06

-0.04

-0.02

0

0.02

0 0.2 0.4 0.6 0.8 1

l i f t c o e f f i c i e n t w . r . t . p a r a m

e t e r p

normalized time

no dynamic pressure con.with dynamic pressure con.

Figure 7.12: Comparison of the sensitivities dC L/dp of the lift coefficient C L w.r.t. perturbationparameter p for the problem with and without dynamic pressure constraint (normalized timescale) for 151 grid points.

-2e-05

-1.5e-05

-1e-05

-5e-06

0

5e-06

0 0.2 0.4 0.6 0.8 1

l i f t c o e f f i c

i e n t w . r . t . p a r a m e t e r e p s

normalized time

no dyn. pressure con.with dyn. pressure con.

Figure 7.13: Comparison of the sensitivities dC L/dε of the lift coefficient C L w.r.t. perturbationparameter ε for the problem with and without dynamic pressure constraint (normalized timescale) for 151 grid points.

c 2006 by M. Gerdts

Page 217: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 217/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 209

-0.4

-0.35

-0.3

-0.25

-0.2

-0.15

-0.1

-0.05

0

0.05

0.1

0 0.2 0.4 0.6 0.8 1

a n g l e o f b a n k w . r . t . p a r a m

e t e r p

normalized time

no dynamic pressure con.with dynamic pressure con.

Figure 7.14: Comparison of the sensitivities dµ/dp of the angle of bank µ w.r.t. perturbationparameter p for the problem with and without dynamic pressure constraint (normalized timescale) for 151 grid points.

-4e-05

-3e-05

-2e-05

-1e-05

0

1e-05

2e-05

0 0.2 0.4 0.6 0.8 1

a n g l e o f b a n k w . r . t . p a r a m e t e r e p

s

normalized time

no dynamic pressure con.with dynamic pressure con.

Figure 7.15: Comparison of the sensitivities dµ/dε of the angle of bank µ w.r.t. perturbationparameter ε for the problem with and without dynamic pressure constraint (normalized timescale) for 151 grid points.

Remark 7.2.4 A corrector iteration method for the reduction of constraint violations which

c 2006 by M. Gerdts

Page 218: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 218/243

210 CHAPTER 7. SELECTED APPLICATIONS AND EXTENSIONS

unavoidedly occur due to the linearization in (3.7.4) was developed in B¨ uskens [B¨ us01]. The application to the emergency problem is discussed in B¨ uskens and Gerdts [BG03] as well.

7.2.4 Numerical Example: Test-Drives

The test-drive to be simulated is again the double-lane change manoeuvre discussed already inSection 7.1.3. Instead of the single-track model a more sophisticated full car model of a BMW1800/2000 is used. The equations of motion form a semi-explicit index 1 DAE (1.2)-(1.2) withnx = 40 differential equations and ny = 4 algebraic equations. The detailed model can be foundin von Heydenaber [vH80] and Gerdts [Ger01a, Ger03b, Ger03c]. We assume that the double-lane change manoeuvre is driven at (almost) constant velocity, i.e. the braking force is set tozero, the acceleration is chosen such that it compensates the effect of rolling resistance, andthe gear is fixed. The remaining control u denotes the steering wheel velocity. There are tworeal-time parameters p1 and p2 involved in the problem. The first one p1 denotes the offset of the test-course with nominal value 3.5 [m], cf. Figure 7.7. The second one p2 denotes the heightof the car’s center of gravity with nominal value 0.56 [m]. While p2 influences the dynamicsof the car, p

1 influences only the state constraints of the optimal control problem. A similar

problem with a different car model was already investigated in Buskens and Gerdts [GB01].

Again, the driver is modeled by formulation of an appropriate optimal control problem withfree final time tf . The car’s initial position on the course is fixed. At final time tf boundaryconditions are given by prescribed x-position (140 [m]) and yaw angle (0 [rad]). In addition, thesteering capability of the driver is restricted by |u(t)| ≤ 3 [rad/s]. The objective is to minimizea linear combination of final time and steering effort, i.e.

40 tf +

tf

t0

u(t)2dt → min .

Figure 7.16 shows the nominal control and the sensitivity differentials for the control at 101 gridpoints obtained with a piecewise linear approximation of the control, i.e. a B-spline representa-tion with k = 2, and the linearized RADAUIIA method of Section 5.4 for time integration.

-1.5

-1

-0.5

0

0.5

1

0 0.2 0.4 0.6 0.8 1

s t e e r i n g

w h e e l v e l o c i t y

normalized time

Figure 7.16: Nominal control for p1 = 3.5 [m] and p2 = 0.56 [m] for 101 grid points.

The nominal objective function value is 273.7460. The optimal final time is tf = 6.7930367 [s]

c 2006 by M. Gerdts

Page 219: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 219/243

7.2. OPEN-LOOP-REAL-TIME CONTROL 211

and its sensitivity differentials are

dtf

dp1= 0.042409,

dtf

dp2= 0.022685.

The second order sufficient conditions in Theorem 3.6.8 are satisfied. The minimal eigenvalueof the reduced Hessian is 0.034. The CPU time for the nominal solution amounts to 318.53 [ s],the overall CPU time including sensitivity analysis is 463.78 [s].The sensitivities are obtained by a sensitivity analysis of the discretized optimal control problemand are depicted in Figures 7.17 and 7.18.

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

0 0.2 0.4 0.6 0.8 1 s t e e r i n g w h e e l v e l o c i t y

w . r . t . p a r a m e t e r 1

normalized time

Figure 7.17: Sensitivity of the steering wheel velocity w.r.t. to the offset in the state constraints

p1 for N = 100 control grid points.

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

0 0.2 0.4 0.6 0.8 1 s t e e r i n g w h e e l

v e l o c i t y w . r . t . p a r a m e t e r 2

normalized time

Figure 7.18: Sensitivity of the steering wheel velocity w.r.t. to the height of the center of gravity p2 for N = 100 control grid points.

Table 7.6 points out the accuracy of the real-time approximation when compared to the optimal

c 2006 by M. Gerdts

Page 220: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 220/243

Page 221: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 221/243

Page 222: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 222/243

Page 223: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 223/243

7.3. DYNAMIC PARAMETER IDENTIFICATION 215

position vector q = (q 1, . . . , q 11) ∈ R11 and the components

q 1 : vertical motion of rear tireq 2 : vertical motion of front tireq 3 : vertical motion of truck chassis

q 4 : rotation about y-axis of truck chassisq 5 : vertical motion of engineq 6 : rotation about y-axis of engineq 7 : vertical motion of driver cabinq 8 : rotation about y-axis of driver cabinq 9 : vertical motion of driver seatq 10 : vertical motion of loading areaq 11 : rotation about y-axis of loading area

The equations of motion are given by the stabilized index 2 DAE

m1q 1 = −F 10 + F 13 − m1g,m2q 2 = −F 20 + F 23 − m2g,

m3q 3 = −F 13 − F 23 + F 35 + F 34 + F 43 + F 53 + F 37 − m3g + λ,

l3q 4 = (a23F 23 − a13F 13 − a37F 37 − a34F 34 − a35F 35 − a43F 43

−a53F 53)cos q 4 − (−az1 cos q 4 + az2 sin q 4)λ,

m4q 5 = −F 43 − F 34 − m4g,

l4q 6 = (b43F 43 − b34F 34)cos q 6,

m5q 7 = −F 53 − F 35 + F 56 − m5g,

l5q 8 = (c53F 53 − c35F 35 − c56F 56)cos q 8,

m6q 9 = −F 56 − m6g,m7q 10 = −F 37 − m7g − λ,

l7q 11 = e37 cos(q 11)F 37 − (ez1 cos q 11 + ez2 sin q 11)λ,

0 = C (q 3, q 4, q 10, q 11, q 3, q 4, q 10, q 11),

0 = c(q 3, q 4, q 10, q 11)

with the algebraic constraint on velocity level (index 2 constraint)

C (q 3, q 4, q 10, q 11, q 3, q 4, q 10, q 11) == −q 3 + q 4(−az1 cos q 4 + az2 sin q 4) + q 10 + q 11(ez1 cos q 11 + ez2 sin q 11)

and the algebraic constraint on position level (stabilizing constraint)

c(q 3, q 4, q 10, q 11) == −q 3 − az1 sin q 4 − az2 cos q 4 + q 10 + ez1 sin q 11 − ez2 cos q 11 + heq.

The model depends on a variety of parameters, cf. Simeon et al. [SGFR94] for a detaileddescription. In this example, we focus in particular on the damping coefficients d 13 and d23, thecoefficient κ influencing the pneumatic spring force law, and the geometric constant heq. Theforces F 13 and F 23 depend on the parameters d13, d23, and κ, whereas the position constraintc depends on heq. Three different parameter identification problems are to be developed in thesequel.

c 2006 by M. Gerdts

Page 224: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 224/243

Page 225: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 225/243

7.3. DYNAMIC PARAMETER IDENTIFICATION 217

optimality and feasibility tolerance within the SQP method was set to 10−7. The objectivefunction in the resulting nonlinear program was scaled by a factor of w = 106. The code DASSLof Petzold [Pet82a] was used for time integration.

In the second identification problem the parameter κ for given d13

and d23

is to be identified outof measured data. In Table 7.8 the results of the second identification problem are summarized.The initial guess for κ was 1.3. The optimality and feasibility tolerance within the SQP methodwas set to 10−7. The objective function in the resulting nonlinear program was not scaled, i.e.w = 1.

Table 7.8: Identification of κ: Dependency of the results with respect to the standard deviationσ of normally distributed measurement errors with expected value 0 [m].

Parameter Nominal σ [m] Result Abs. Error Rel. Error CPU [s]κ 1.4 10−3 1.4129911 1.29911 · 10−2 9.28 · 10−3 12.36

10−4 1.4012414 1.2414 · 10−3 8.87 · 10−4 12.2110−5 1.4001233 1.233 · 10−4 8.81 · 10−5 12.2410−6 1.4000121 1.21 · 10−5 8.64 · 10−6 12.21

The parameter heq influences the algebraic constraint on position level. Hence, a third identifi-cation problem is given by

(iii) The parameter heq as well as the unknown initial vertical positions q 3(t0) and q 10(t0) areto be identified out of the measured data for given d13, d23 and κ.

This problem is more complicated than the previous problems, since the sought values occurin the stabilizing position constraint. During the optimization process the current iterates ingeneral are inconsistent with the position constraint. Hence, the projection method describedin Section 6.1.1 is essentially needed to compute consistent values.

Table 7.9 summarizes the results of the third identification problem. As before the optimalityand feasibility tolerance within the SQP method was set to 10−7. The initial guesses for heq,q 3(t0) and q 10(t0) were 0.7 [m], 0 [m] and 0 [m], respectively. The objective function in theresulting nonlinear program was scaled with w = 2. A multiple shooting approach similar to thereduced approach was used with three equidistant multiple shooting nodes. Consistent valuesat the multiple shooting nodes are obtained by the projection method as in Section 6.1.1.

c 2006 by M. Gerdts

Page 226: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 226/243

Page 227: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 227/243

Bibliography

[AF96] Adjiman, C. S. and Floudas, C. A. Rigorous convex underestimators for general twice-differentiable problems . Journal of Global Optimization, 9 (1); 23–40, 1996.

[AF01] Adjiman, C. S. and Floudas, C. A. The αBB global optimization algorithm for nonconvex problems: An overview . In From local to global optimization. Papers

from the conference dedicated to Professor Hoang Tuy on the occasion of his 70th birthday, Rimforsa, Sweden, August 1997 (A. M. et al., editor), volume 53 of Non-convex Optimization and Applications , pp. 155–186. Kluwer Academic Publishers,Dordrecht, 2001.

[AF04a] Akrotirianakis, I. G. and Floudas, C. A. Computational experience with a new class of convex underestimators: Box-constrained NLP problems . Journal of GlobalOptimization, 29 (3); 249–264, 2004.

[AF04b] Akrotirianakis, I. G. and Floudas, C. A. A new class of improved convex underes-timators for twice continuously differentiable constrained NLPs . Journal of GlobalOptimization, 30 (4); 367–390, 2004.

[Alt91] Alt, W. Sequential Quadratic Programming in Banach Spaces . In Advances in Optimization (W. Oettli and D. Pallaschke, editors), pp. 281–301. Springer, Berlin,1991.

[Alt02] Alt, W. Nichtlineare Optimierung: Eine Einf¨ uhrung in Theorie, Verfahren und Anwendungen . Vieweg, Braunschweig/Wiesbaden, 2002.

[AM98] Arnold, M. and Murua, A. Non-stiff integrators for differential-algebraic systems of index 2 . Numer. Algorithms, 19 (1-4); 25–41, 1998.

[AM01] Augustin, D. and Maurer, H. Computational sensitivity analysis for state con-strained optimal control problems . Annals of Operations Research, 101; 75–99,2001.

[AMF95] Androulakis, I., Maranas, C. and Floudas, C. αBB: A global optimization method

for general constrained nonconvex problems . Journal of Global Optimization, 7 (4);337–363, 1995.

[AP91] Ascher, U. M. and Petzold, L. R. Projected implicit Runge-Kutta methods for differential-algebraic equations . SIAM Journal on Numerical Analysis, 28 (4); 1097–1120, 1991.

[Arn95] Arnold, M. A perturbation analysis for the dynamical simulation of mechnical multi-body systems . Applied Numerical Mathematics, 18 (1); 37–56, 1995.

[Arn98] Arnold, M. Half-explicit Runge-Kutta methods with explicit stages for differential-algebraic systems of index 2 . BIT, 38 (3); 415–438, 1998.

219

Page 228: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 228/243

220 BIBLIOGRAPHY

[Bac06] Backes, A. Extremalbedingungen f¨ ur Optimierungs-Probleme mit Algebro-Differentialgleichungen . Ph.D. thesis, Mathematisch-NaturwissenschaftlicheFakultat, Humboldt-Universitat Berlin, Berlin, Germany, 2006.

[BCP96] Brenan, K. E., Campbell, S. L. and Petzold, L. R. Numerical Solution of Initial-

Value Problems in Differential-Algebraic Equations , volume 14 of Classics In Ap-plied Mathematics . SIAM, Philadelphia, 1996.

[BE88] Brenan, K. E. and Engquist, B. E. Backward Differentiation Approximations of Nonlinear Differential/Algebraic Systems . Mathematics of Computations, 51 (184);659–676, 1988.

[Bet90] Betts, J. T. Sparse Jacobian Updates in the Collocation Method for Optimal Control Problems . Journal of Guidance, Control and Dynamics, 13 (3); 409–415, 1990.

[BF97] Billups, S. C. and Ferris, M. C. QPCOMP: A quadratic programming based solver for mixed complementarity problems . Mathematical Programming, 76 (3); 533–562,

1997.

[BG00] Buskens, C. and Gerdts, M. Numerical Solution of Optimal Control Problems with DAE Systems of Higher Index . In Optimalsteuerungsprobleme in der Luft- und Raumfahrt, Workshop in Greifswald des Sonderforschungsbereichs 255: Transat-mosp¨ arische Flugsysteme , pp. 27–38. Munchen, 2000.

[BG03] Buskens, C. and Gerdts, M. Emergency Landing of a Hypersonic Flight System:A Corrector Iteration Method for Admissible Real–Time Optimal Control Approx-imations . In Optimalsteuerungsprobleme in der Luft- und Raumfahrt, Workshopin Greifswald des Sonderforschungsbereichs 255: Transatmosp¨ arische Flugsysteme ,

pp. 51–60. Munchen, 2003.[BG05] Buskens, C. and Gerdts, M. Differentiability of Consistency Functions for DAE

Systems . Journal of Optimization Theory and Applications, 125 (1); 37–61, 2005.

[BGK+83] Bank, B., Guddat, J., Klatte, D., Kummer, B. and Tammer, K. Non-linear para-metric optimization . Birkhauser, Basel, 1983.

[BGR97] Barcley, A., Gill, P. E. and Rosen, J. B. SQP methods and their application tonumerical optimal control . Report NA 97-3, Dep. of Mathematics, University of California, San Diego, 1997.

[BH75] Bryson, A. E. and Ho, Y.-C. Applied Optimal Control . Hemisphere PublishingCorporation, Washington, 1975.

[BH92] Betts, J. T. and Huffman, W. P. Application of Sparse Nonlinear Programming to Trajectory Optimization . Journal of Guidance, Control and Dynamics, 15 (1);198–206, 1992.

[BH99] Betts, J. T. and Huffman, W. P. Exploiting Sparsity in the Direct Transcription Method for Optimal Control . Computational Optimization and Applications, 14 (2);179–201, 1999.

[BM96] Buskens, C. and Maurer, H. Sensitivity Analysis and Real-Time Control of Nonlin-ear Optimal Control Systems via Nonlinear Programming Methods . Proceedings of

c 2006 by M. Gerdts

Page 229: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 229/243

BIBLIOGRAPHY 221

the 12th Conference on Calculus of Variations, Optimal Control and Applications.Trassenheide, 1996.

[BM01a] Buskens, C. and Maurer, H. Sensitivity Analysis and Real-Time Control of Paramet-ric Optimal Control Problems Using Nonlinear Programming Methods . In Online

Optimization of Large Scale Systems (M. Grotschel, S. O. Krumke and J. Rambau,editors), pp. 56–68. Springer, 2001.

[BM01b] Buskens, C. and Maurer, H. Sensitivity Analysis and Real-Time Optimization of Parametric Nonlinear Programming Problems . In Online Optimization of Large Scale Systems (M. Grotschel, S. O. Krumke and J. Rambau, editors), pp. 3–16.Springer, 2001.

[BO75] Blum, E. and Oettli, W. Mathematische Optimierung . volume 20 of ¨ Okonometrie und Unternehmensforschung . Springer-Verlag Berlin Heidelberg New York, Berlin,1975.

[Boc87] Bock, H. G. Randwertproblemmethoden zur Parameteridentifizierung in Syste-men nichtlinearer Differentialgleichungen . volume 183 of Bonner Mathematische Schriften . Bonn, 1987.

[BP84] Bock, H. G. and Plitt, K. J. A Multiple Shooting Algorithm for Direct Solution of Optimal Control Problems . Proceedings of the 9th IFAC Worldcongress, Budapest,Hungary. 1984.

[BP89] Brenan, K. E. and Petzold, L. R. The numerical solution of higher index differen-tial/algebraic equations by implicit methods . SIAM Journal on Numerical Analysis,26 (4); 976–996, 1989.

[BS79] Bazaraa, M. S. and Shetty, C. M. Nonlinear Programming: Theory and Algorithms .John Wiley & Sons, 1979.

[BS00] Bonnans, J. F. and Shapiro, A. Perturbation Analysis of Optimization Problems .Springer Series in Operations Research. Springer, New York, 2000.

[BSS93] Bazaraa, M. S., Sherali, H. D. and Shetty, C. M. Nonlinear Programming: Theory and Algorithms . John Wiley & Sons, 2nd edition, 1993.

[BTW82] Boggs, P. T., Tolle, J. W. and Wang, P. On the local convergence of quasi-newton methods for constrained optimization . SIAM Journal on Control and Optimization,20 (2); 161–171, 1982.

[Bul71] Bulirsch, R. Die Mehrzielmethode zur numerischen L¨ osung von nichtlinearen Randwertproblemen und Aufgaben der optimalen Steuerung . Report der Carl-Cranz-Gesellschaft, 1971.

[Bus98] Buskens, C. Optimierungsmethoden und Sensitivit¨ atsanalyse f¨ ur optimale Steuer-prozesse mit Steuer- und Zustandsbeschr¨ ankungen . Ph.D. thesis, Fachbereich Math-ematik, Westfalische Wilhems-Universitat Munster, 1998.

[Bus01] Buskens, C. Real-Time Solutions for Perturbed Optimal Control Problems by a Mixed Open- and Closed-Loop Strategy . In Online Optimization of Large Scale Sys-tems (M. Grotschel, S. O. Krumke and J. Rambau, editors), pp. 105–116. Springer,2001.

c 2006 by M. Gerdts

Page 230: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 230/243

222 BIBLIOGRAPHY

[CG95] Campbell, S. L. and Gear, C. W. The index of general nonlinear DAEs . NumerischeMathematik, 72; 173–196, 1995.

[CH52] Curtiss, C. F. and Hirschfelder, J. O. Integration of stiff equations . Proceedingsof the National Academy of Sciences of the United States of America, 38; 235–243,

1952.

[Chu94] Chudej, K. Optimale Steuerung des Aufstiegs eines zweistufigen Hyperschall-Raumtransporters . Ph.D. thesis, Mathematisches Institut, Technische UniversitatMunchen, 1994.

[CL82] Chernousko, F. L. and Lyubushin, A. A. Method of successive approximations for solution of optimal control problems . Optimal Control Applications and Methods,3; 101–114, 1982.

[Cla83] Clarke, F. H. Optimization and Nonsmooth Analysis . John Wiley & Sons, NewYork, 1983.

[CLPS03] Cao, Y., Li, S., Petzold, L. R. and Serban, R. Adjoint sensitivity analysis for differential-algebraic equations: The adjoint DAE system and its numerical solution .SIAM Journal on Scientific Computing, 24 (3); 1076–1089, 2003.

[CS85] Caracotsios, M. and Stewart, W. E. Sensitivity analysis of initial-boundary-value problems with mixed PDEs and algebraic equations . Computers chem. Engng, 19 (9);1019–1030, 1985.

[DB02] Deuflhard, P. and Bornemann, F. Scientific Computing with Ordinary Differential Equations . volume 42 of Texts in Applied Mathematics . Springer-Verlag New York,New York, 2002.

[Deu74] Deuflhard, P. A modified Newton method for the solution of ill-conditioned systems of nonlinear equations with apllication to multiple shooting . Numerische Mathe-matik, 22; 289–315, 1974.

[Deu79] Deuflhard, P. A Stepsize Control for Continuation Methods and its Special Ap-plication to Multiple Shooting Techniques . Numerische Mathematik, 33; 115–146,1979.

[DG86a] Duff, I. S. and Gear, C. W. Computing the structural index . SIAM Journal onAlgebraic Discrete Methods, 7 (4); 594–603, 1986.

[DG86b] Duran, M. A. and Grossmann, I. E. An outer-approximation algorithm for a class of mixed-integer nonlinear programs . Mathematical Programming, 36; 307–339, 1986.

[DH91] Deuflhard, P. and Hohmann, A. Numerische Mathematik . de Gruyter, Berlin, 1991.

[DHM00] Dontchev, A. L., Hager, W. W. and Malanowski, K. Error Bounds for Euler Approx-imation of a State and Control Constrained Optimal Control Problem . NumericalFunctional Analysis and Optimization, 21 (5 & 6); 653–682, 2000.

[DHV00] Dontchev, A. L., Hager, W. W. and Veliov, V. M. Second-Order Runge-Kutta Ap-proximations in Control Constrained Optimal Control . SIAM Journal on NumericalAnalysis, 38 (1); 202–226, 2000.

c 2006 by M. Gerdts

Page 231: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 231/243

BIBLIOGRAPHY 223

[DL99] Devdariani, E. N. and Ledyaev, Y. S. Maximum Principle for Implicit Control Systems . Applied Mathematics and Optimization, 40; 79–103, 1999.

[DLO+77] Diekhoff, H.-J., Lory, P., Oberle, H., Pesch, H.-J., Rentrop, P. and Seydel, R.Comparing routines for the numerical solution of initial value problems of ordinary

differential equations in multiple shooting . Numerische Mathematik, 27; 449–469,1977.

[DPR76] Deuflhard, P., Pesch, H. J. and Rentrop, P. A modified continuation method for the numerical solution of nonlinear two-point boundary value problems by shooting techniques . Numerische Mathematik, 26; 327–343, 1976.

[dPV97] de Pinho, M. and Vinter, R. B. Necessary Conditions for Optimal Control Prob-lems Involving Nonlinear Differential Algebraic Equations . Journal of MathematicalAnalysis and Applications, 212; 493–516, 1997.

[DW02] Dingguo, P. and Weiwen, T. Globally convergent inexact generalized Newton’s meth-

ods for nonsmooth equations . Journal of Computational and Applied Mathematics,138; 37–49, 2002.

[Eic93] Eich, E. Convergence results for a coordinate projection method applied to mechani-cal systems with algebraic constraints . SIAM Journal on Numerical Analysis, 30 (5);1467–1482, 1993.

[EKKvS99] Engl, G., Kroner, A., Kronseder, T. and von Stryk, O. Numerical Simulation and Optimal Control of Air Separation Plants . In High Performance Scientific and Engineering Computing (H.-J. Bungartz, F. Durst and C. Zenger, editors), volume 8of Lecture Notes in Computational Science and Engineering , pp. 221–231. Springer,1999.

[Fia83] Fiacco, A. V. Introduction to Sensitivity and Stability Analysis in Nonlinear Pro-gramming , volume 165 of Mathematics in Science and Engineering . Academic Press,New York, 1983.

[Fis97] Fischer, A. Solution of monotone complementarity problems with locally Lipschitzian functions . Mathematical Programming, 76 (3); 513–532, 1997.

[FK97] Facchinei, F. and Kanzow, C. A nonsmooth inexact Newton method for the solution of large-scale nonlinear complementarity problems . Mathematical Programming,76 (3); 493–512, 1997.

[FL91] Fuhrer, C. and Leimkuhler, B. J. Numerical solution of differential-algebraic equa-tions for constraint mechanical motion . Numerische Mathematik, 59; 55–69, 1991.

[FL94] Fletcher, R. and Leyffer, S. Solving mixed integer nonlinear programs by outer approximation . Mathematical Programming, 66; 327–349, 1994.

[FM90] Fiacco, A. V. and McCormick, G. P. Nonlinear Programming: Sequential Uncon-strained Minimization Techniques , volume 4 of Classics In Applied Mathematics .SIAM, Philadelphia, 1990.

[FTB97] Feehery, W. F., Tolsma, J. E. and Barton, P. I. Efficient sensitivity analysis of large-scale differential-algebraic systems . Applied Numerical Mathematics, 25; 41–54, 1997.

c 2006 by M. Gerdts

Page 232: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 232/243

224 BIBLIOGRAPHY

[Fuh88] Fuhrer, C. Differential-algebraische Gleichungssysteme in mechanischen Mehrk¨ orpersystemen: Theorie, numerische Ans¨ atze und Anwendungen . Ph.D. the-sis, Fakultat fur Mathematik und Informatik, Technische Universitat Munchen,1988.

[GB01] Gerdts, M. and Buskens, C. Computation of Consistent Initial Values for Optimal Control Problems with DAE Systems of Higher Index . ZAMM, 81 S2; 249–250,2001.

[GB02] Gerdts, M. and Buskens, C. Consistent Initialization of Sensitivity Matrices for a Class of Parametric DAE Systems . BIT Numerical Mathematics, 42 (4); 796–813,2002.

[Gea71] Gear, C. W. Simultaneous Numerical Solution of Differential-Algebraic Equations .IEEE Transactions on Circuit Theory, 18 (1); 89–95, 1971.

[Gea88] Gear, C. W. Differential-algebraic equation index transformations . SIAM Journal

on Scientific and Statistical Computing, 9; 39–47, 1988.

[Gea90] Gear, C. W. Differential algebraic equations, indices, and integral algebraic equa-tions . SIAM Journal on Numerical Analysis, 27 (6); 1527–1534, 1990.

[Ger01a] Gerdts, M. Numerische Methoden optimaler Steuerprozesse mit differential-algebraischen Gleichungssystemen h¨ oheren Indexes und ihre Anwendungen in der Kraftfahrzeugsimulation und Mechanik . volume 61 of Bayreuther Mathematische Schriften . Bayreuth, 2001.

[Ger01b] Gerdts, M. SODAS – Software for Optimal Control Problems with Differential-Algebraic Systems: User’s guide . Technical report, Institut fur Mathematik, Uni-

versitat Bayreuth, 2001.

[Ger03a] Gerdts, M. Direct Shooting Method for the Numerical Solution of Higher Index DAE Optimal Control Problems . Journal of Optimization Theory and Applications,117 (2); 267–294, 2003.

[Ger03b] Gerdts, M. A Moving Horizon Technique for the Simulation of Automobile Test-Drives . ZAMM, 83 (3); 147–162, 2003.

[Ger03c] Gerdts, M. Optimal Control and Real-Time Optimization of Mechanical Multi-Body Systems . ZAMM, 83 (10); 705–719, 2003.

[Ger04] Gerdts, M. Parameter Optimization in Mechanical Multibody Systems and Lin-earized Runge-Kutta Methods . In Progress in Industrial Mathematics at ECMI 2002 (A. Buikis, R. Ciegis and A. D. Flitt, editors), volume 5 of Mathematics in Industry , pp. 121–126. Springer, 2004.

[Ger05a] Gerdts, M. Local minimum principle for optimal control problems sub- ject to index one differential-algebraic equations . Technical report, De-partment of Mathematics, University of Hamburg, http://www.math.uni-hamburg.de/home/gerdts/Report index1.pdf, 2005.

[Ger05b] Gerdts, M. Numerische L¨ osungsverfahren f¨ ur Optimalsteuerungsprobleme . LectureNote, University of Hamburg, Department of Mathematics, 2005.

c 2006 by M. Gerdts

Page 233: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 233/243

BIBLIOGRAPHY 225

[Ger05c] Gerdts, M. On the Convergence of Linearized Implicit Runge-Kutta Methods and their Use in Parameter Optimization . Mathematica Balkanica (New Series), 19;75–83, 2005.

[Ger05d] Gerdts, M. Parameter Identification in Higher Index DAE Systems . Technical

report, Department of Mathematics, Universitat Hamburg, 2005.

[Ger05e] Gerdts, M. Solving Mixed-Integer Optimal Control Problems by Branch&Bound:A Case Study from Automobile Test-Driving with Gear Shift . Optimal Control,Applications and Methods, 26 (1); 1–18, 2005.

[Ger06] Gerdts, M. A variable time transformation method for mixed-integer optimal control problems . Optimal Control, Applications and Methods, 27 (3); 169–182, 2006.

[GI83] Goldfarb, D. and Idnani, A. A numerically stable dual method for solving strictly convex quadratic programs . Mathematical Programming, 27; 1–33, 1983.

[Gir72] Girsanov, I. V. Lectures on Mathematical Theory of Extremum Problems . vol-ume 67 of Lecture Notes in Economics and Mathematical Systems . Springer, Berlin-Heidelberg-New York, 1972.

[GK97] Grossmann, I. and Kravanja, Z. Mixed-Integer Nonlinear Programming: A Survey of Algorithms and Applications . volume 93 of The IMA Volumes in Mathematics and its Applications , pp. 73–100. Springer, New York,Berlin,Heidelberg, 1997.

[GK99] Geiger, C. and Kanzow, C. Numerische Verfahren zur L¨ osung unrestringierter Optimierungsaufgaben . Springer, Berlin-Heidelberg-New York, 1999.

[GK02] Geiger, C. and Kanzow, C. Theorie und Numerik restringierter Optimierungsauf-gaben . Springer, Berlin-Heidelberg-New York, 2002.

[GLG85] Gear, C. W., Leimkuhler, B. and Gupta, G. K. Automatic integration of Euler-Lagrange equations with constraints . Journal of Computational and Applied Math-ematics, 12/13; 77–90, 1985.

[GM78] Gill, P. E. and Murray, W. Numerically stable methods for quadratic programming .Mathematical Programming, 14; 349–372, 1978.

[GMS94] Gill, P. E., Murray, W. and Saunders, M. A. Large-scale SQP Methods and their Application in Trajectory Optimization , volume 115 of International Series of Nu-

merical Mathematics , pp. 29–42. Birkhauser, Basel, 1994.

[GMS02] Gill, P. E., Murray, W. and Saunders, M. A. SNOPT: An SQP algorithm for large-scale constrained optimization . SIAM Journal on Optimization, 12 (4); 979–1006,2002.

[GMSW91] Gill, P. E., Murray, W., Saunders, M. A. and Wright, M. H. Inertia-controlling methods for general quadratic programming . SIAM Review, 33 (1); 1–36, 1991.

[GMSW98] Gill, P. E., Murray, W., Saunders, M. A. and Wright, M. H. User’s guide for NPSOL5.0: A FORTRAN package for nonlinear programming . Technical Report NA 98-2,Department of Mathematics, University of California, San Diego,California, 1998.

c 2006 by M. Gerdts

Page 234: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 234/243

226 BIBLIOGRAPHY

[GMW81] Gill, P. E., Murray, W. and Wright, M. H. Practical Optimization . Academic Press,London, 1981.

[GP84] Gear, C. W. and Petzold, L. ODE methods for the solution of differential/algebraic systems . SIAM Journal on Numerical Analysis, 21 (4); 716–728, 1984.

[GR80] Gopfert, A. and Riedrich, T. Funktionalanalysis . Teubner, Leipzig, 1980.

[Gri00] Griewank, A. Evaluating derivatives. Principles and techniques of algorithmic dif- ferentiation , volume 19 of Frontiers in Applied Mathematics . SIAM, Philadelphia,2000.

[Gri03] Griewank, A. A mathematical view of automatic differentiation . Acta Numerica,12; 321–398, 2003.

[Gru96] Grupp, F. Parameteridentifizierung nichtlinearer mechanischer Deskriptorsysteme mit Anwendungen in der Rad–Schiene–Dynamik . Fortschritt-Berichte VDI Reihe

8, Nr. 550. VDI–Verlag, Dusseldorf, 1996.

[GS01] Ghildyal, V. and Sahinidis, N. V. Solving global optimization problems with Baron .In From local to global optimization. Papers from the conference dedicated to Pro-

fessor Hoang Tuy on the occasion of his 70th birthday, Rimforsa, Sweden, August 1997 (A. M. et al., editor), volume 53 of Nonconvex Optimization and Applications ,pp. 205–230. Kluwer Academic Publishers, Dordrecht, 2001.

[Gun95] Gunther, M. Ladungsorientierte Rosenbrock-Wanner-Methoden zur numerischen Simulation digitaler Schaltungen , volume 168 of VDI Fortschrittberichte Reihe 20:Rechnergest¨ utzte Verfahren . VDI-Verlag, 1995.

[Hag00] Hager, W. W. Runge-Kutta methods in optimal control and the transformed adjoint system . Numerische Mathematik, 87 (2); 247–282, 2000.

[Han77] Han, S. P. A Globally Convergent Method for Nonlinear Programming . Journal of Optimization Theory and Applications, 22 (3); 297–309, 1977.

[HCB93] Hiltmann, P., Chudej, K. and Breitner, M. H. Eine modifizierte Mehrziel-methode zur L¨ osung von Mehrpunkt-Randwertproblemen . Technical Report 14,Sonderforschungsbereich 255 der Deutschen Forschungsgemeinschaft: Transatmo-spharische Flugsysteme, Lehrstuhl fur Hohere und Numerische Mathematik, Tech-nische Universitat Munchen, 1993.

[Hei92] Heim, A. Parameteridentifizierung in differential-algebraischen Gleichungssyste-men . Master’s thesis, Mathematisches Institut, Technische Universitat Munchen,1992.

[Hes66] Hestenes, M. R. Calculus of variations and optimal control theory . John Wiley &Sons, New York, 1966.

[Hin97] Hinsberger, H. Ein direktes Mehrzielverfahren zur L¨ osung von Optimals-teuerungsproblemen mit großen, differential-algebraischen Gleichungssystemen und Anwendungen aus der Verfahrenstechnik . Ph.D. thesis, Institut fur Mathematik,Technische Universitat Clausthal, 1997.

c 2006 by M. Gerdts

Page 235: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 235/243

BIBLIOGRAPHY 227

[HL69] Hermes, H. and Lasalle, J. P. Functional Analysis and Time Optimal Control ,volume 56 of Mathematics in Science and Engineering . Academic Press, New York,1969.

[HLR89] Hairer, E., Lubich, C. and Roche, M. The numerical solution of differential-algebraic systems by Runge-Kutta methods . volume 1409 of Lecture Notes in Mathematics .Springer, Berlin-Heidelberg-New York, 1989.

[HP87] Hargraves, C. R. and Paris, S. W. Direct trajectory optimization using nonlinear programming and collocation . Journal of Guidance, Control and Dynamics, 10;338–342, 1987.

[HPR92] Han, S. P., Pang, J. S. and Rangaraj, N. Globally convergent Newton methods for nonsmooth equations . Mathematics of Operations Research, 17 (3); 586–607, 1992.

[HSV95] Hartl, R. F., Sethi, S. P. and Vickson, G. A Survey of the Maximum Principles for

Optimal Control Problems with State Constraints . SIAM Review, 37 (2); 181–218,1995.

[HW96] Hairer, E. and Wanner, G. Solving ordinary differential equations II: Stiff and differential-algebraic problems , volume 14. Springer Series in Computational Math-ematics, Berlin-Heidelberg-New York, 2nd edition, 1996.

[IT79] Ioffe, A. D. and Tihomirov, V. M. Theory of extremal problems . volume 6 of Studies in Mathematics and its Applications . North-Holland Publishing Company,Amsterdam, New York, Oxford, 1979.

[Jay93] Jay, L. Collocation methods for differential-algebraic equations of index 3 . Nu-merische Mathematik, 65; 407–421, 1993.

[Jay95] Jay, L. Convergence of Runge-Kutta methods for differential-algebraic systems of index 3 . Applied Numerical Mathematics, 17; 97–118, 1995.

[Jia99] Jiang, H. Global convergence analysis of the generalized Newton and Gauss-Newton methods of the Fischer-Burmeister equation for the complementarity problem . Math-ematics of Operations Research, 24 (3); 529–543, 1999.

[JLS71] Jacobson, D. H., Lele, M. M. and Speyer, J. L. New Necessary Conditions of Optimality for Constrained Problems with State-Variable Inequality Constraints .

Journal of Mathematical Analysis and Applications, 35; 255–284, 1971.

[JQ97] Jiang, H. and Qi, L. A new nonsmooth equations approach to nonlinear comple-mentarity problems . SIAM Journal on Control and Optimization, 35 (1); 178–193,1997.

[KE85] Kramer-Eis, P. Ein Mehrzielverfahren zur numerischen Berechnung optimaler Feedback-Steuerungen bei beschr¨ ankten nichtlinearen Steuerungsproblemen . volume166 of Bonner Mathematische Schriften . Bonn, 1985.

[Kie98] Kiehl, M. Sensitivity Analysis of ODEs and DAEs - Theory and Implementation Guide . Technische Universitat Munchen, TUM-M5004, 1998.

c 2006 by M. Gerdts

Page 236: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 236/243

228 BIBLIOGRAPHY

[Kla90] Klatte, D. Nonlinear optimization problems under data perturbations . volume 378of Lecture Notes in Economics and Mathematical Systems , pp. 204–235. Springer,Berlin-Heidelberg-New York, 1990.

[KM97] Kunkel, P. and Mehrmann, V. The linear quadratic optimal control problem for

linear descriptor systems with variable coefficients . Mathematics of Control, Signals,and Systems, 10 (3); 247–264, 1997.

[KM04] Kurina, G. A. and Marz, R. On linear-quadratic optimal control problems for time-varying descriptor systems . SIAM Journal on Control and Optimization, 42 (6);2062–2077, 2004.

[Kno75] Knobloch, H. W. Das Pontryaginsche Maximumprinzip f¨ ur Probleme mit Zustands-beschr¨ ankungen I und II . Zeitschrift fur Angewandte Mathematik und Mechanik,55; 545–556, 621–634, 1975.

[KOS77] Kufner, A., Oldrich, J. and Svatopluk, F. Function Spaces . Noordhoff International

Publishing, Leyden, 1977.

[Kow69] Kowalsky, H.-J. Lineare Algebra . Walter de Gruyter & Co, Berlin, 4th edition,1969.

[Kra88] Kraft, D. A Software Package for Sequential Quadratic Programming . DFVLR-FB-88-28, Oberpfaffenhofen, 1988.

[Kre82] Kreindler, E. Additional Necessary Conditions for Optimal Control with State-Variable Inequality Constraints . Journal of Optimization Theory and Applications,38 (2); 241–250, 1982.

[Kur76] Kurcyusz, S. On the Existence and Nonexistence of Lagrange Multipliers in Banach Spaces . Journal of Optimization Theory and Applications, 20 (1); 81–110, 1976.

[KWW78] Kirsch, A., Warth, W. and Werner, J. Notwendige Optimalit¨ atsbedingungen und ihre Anwendung . volume 152 of Lecture Notes in Economics and Mathematical Systems .Springer, Berlin-Heidelberg-New York, 1978.

[Lei95] Leineweber, D. B. Analyse und Restrukturierung eines Verfahrens zur direkten L¨ osung von Optimal-Steuerungsproblemen . Master’s thesis, Interdisziplinares Zen-trum fur Wissenschaftliches Rechnen, Universitat Heidelberg, 1995.

[Lem62] Lemke, C. E. A method of solution for quadratic programs . Management Science,

8; 442–453, 1962.

[Lem71a] Lempio, F. Lineare Optimierung in unendlichdimensionalen Vektorr¨ aumen . Com-puting, 8; 284–290, 1971.

[Lem71b] Lempio, F. Separation und Optimierung in linearen R¨ aumen . Ph.D. thesis, Univer-sitat Hamburg, Hamburg, 1971.

[Lem72] Lempio, F. Tangentialmannigfaltigkeiten und infinite Optimierung . Habilitationss-chrift, Universitat Hamburg, Hamburg, 1972.

[Ley01] Leyffer, S. Integrating SQP and Branch-and-Bound for Mixed Integer Nonlinear Programming . Computational Optimization and Applications, 18; 295–309, 2001.

c 2006 by M. Gerdts

Page 237: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 237/243

BIBLIOGRAPHY 229

[LM80] Lempio, F. and Maurer, H. Differential stability in infinite-dimensional nonlinear programming . Applied Mathematics and Optimization, 6; 139–152, 1980.

[LP86] Lotstedt, P. and Petzold, L. R. Numerical Solution of Nonlinear Differential Equa-tions with Algebraic Constraints I: Convergence Results for Backward Differentia-tion Formulas . Mathematics of Computation, 46; 491–516, 1986.

[LPG91] Leimkuhler, B., Petzold, L. R. and Gear, C. W. Approximation methods for the con-sistent initialization of differential-algebraic equations . SIAM Journal on NumericalAnalysis, 28 (1); 205–226, 1991.

[LS76] Ljusternik, L. A. and Sobolew, W. I. Elemente der Funktionalanalysis . Verlag HarriDeutsch, Zurich-Frankfurt/Main-Thun, 1976.

[LTC98] Lee, H. W. J., Teo, K. L. and Cai, X. Q. An Optimal Control Approach to Nonlinear Mixed Integer Programming Problems . Computers & Mathematics with Applica-

tions, 36 (3); 87–105, 1998.

[LTRJ97] Lee, H. W. J., Teo, K. L., Rehbock, V. and Jennings, L. S. Control parameteriza-tion enhancing technique for time optimal control problems . Dynamic Systems andApplications, 6 (2); 243–262, 1997.

[LTRJ99] Lee, H. W. J., Teo, K. L., Rehbock, V. and Jennings, L. S. Control parametriza-tion enhancing technique for optimal discrete-valued control problems . Automatica,35 (8); 1401–1407, 1999.

[Lue69] Luenberger, D. G. Optimization by Vector Space Methods . John Wiley & Sons,New York-London-Sydney-Toronto, 1969.

[LVBB+04] Laurent-Varin, J., Bonnans, F., Berend, N., Talbot, C. and Haddou, M. On the refinement of discretization for optimal control problems . IFAC Symposium onAutomatic Control in Aerospace, St. Petersburg, 2004.

[MA01] Maurer, H. and Augustin, D. Sensitivity Analysis and Real-Time Control of Para-metric Optimal Control Problems Using Boundary Value Methods . In Online Op-timization of Large Scale Systems (M. Grotschel, S. O. Krumke and J. Rambau,editors), pp. 17–55. Springer, 2001.

[Mac88] Machielsen, K. C. P. Numerical Solution of Optimal Control Problems with State

Constraints by Sequential Quadratic Programming in Function Space . volume 53 of CWI Tract . Centrum voor Wiskunde en Informatica, Amsterdam, 1988.

[Mal97] Malanowski, K. Sufficient Optimality Conditions for Optimal Control subject toState Constraints . SIAM Journal on Control and Optimization, 35 (1); 205–227,1997.

[Man94] Mangasarian, O. L. Nonlinear Programming , volume 10 of Classics In Applied Mathematics . SIAM, Philadelphia, 1994.

[Mar95] Marz, R. On linear differential-algebraic equations and linerizations . Applied Nu-merical Mathematics, 18 (1); 267–292, 1995.

c 2006 by M. Gerdts

Page 238: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 238/243

230 BIBLIOGRAPHY

[Mar98a] Marz, R. Criteria of the trivial solution of differential algebraic equations with small nonlinearities to be asymptotically stable . Journal of Mathematical Analysisand Applications, 225 (2); 587–607, 1998.

[Mar98b] Marz, R. EXTRA-ordinary differential equations: Attempts to an analysis of

differential-algebraic systems . In European congress of mathematics (A. Balog, ed-itor), volume 1 of Prog. Math. 168 , pp. 313–334. Birkhauser, Basel, 1998.

[Mau77] Maurer, H. On Optimal Control Problems with Boundary State Variables and Con-trol Appearing Linearly . SIAM Journal on Control and Optimization, 15 (3); 345–362, 1977.

[Mau79] Maurer, H. On the Minimum Principle for Optimal Control Problems with State Constraints . Schriftenreihe des Rechenzentrums der Universitat Munster, 41, 1979.

[Mau81] Maurer, H. First and Second Order Sufficient Optimality Conditions in Mathemat-

ical Programming and Optimal Control . Mathematical Programming Study, 14;163–177, 1981.

[May91] Mayr, R. Verfahren zur Bahnfolgeregelung f¨ ur ein automatisch gef¨ uhrtes Fahrzeug .Ph.D. thesis, Fakultat fur Elektrotechnik, Universitat Dortmund, 1991.

[MBM97] Malanowski, K., Buskens, C. and Maurer, H. Convergence of Approximations toNonlinear Optimal Control Problems . In Mathematical programming with data per-turbations (A. Fiacco, editor), volume 195, pp. 253–284. Dekker. Lecture Notes inPure and Applied Mathematics, 1997.

[Meh91] Mehrmann, V. The autonomous linear quadratic control problem. Theory and nu-

merical solution . volume 163 of Lecture Notes in Control and Information Sciences .Springer, Berlin, 1991.

[MF05] Meyer, C. A. and Floudas, C. A. Convex underestimation of twice continuously differentiable functions by piecewise quadratic perturbation: Spline α BB underesti-mators . Journal of Global Optimization, 32 (2); 221–258, 2005.

[Mit90] Mitschke, M. Dynamik der Kraftfahrzeuge, Band C: Fahrverhalten . Springer,Berlin-Heidelberg-New York, 2nd edition, 1990.

[MM96] Malanowski, K. and Maurer, H. Sensitivity analysis for parametric control problems with control-state constraints . Computational Optimization and Applications, 5 (3);

253–283, 1996.

[MM98] Malanowski, K. and Maurer, H. Sensitivity analysis for state constrained optimal control problems . Discrete and Continuous Dynamical Systems, 4 (2); 3–14, 1998.

[MM01] Malanowski, K. and Maurer, H. Sensitivity analysis for optimal control problems subject to higher order state constraints . Annals of Operations Research, 101; 43–73,2001.

[MMP04] Malanowski, K., Maurer, H. and Pickenhain, S. Second-order Sufficient Conditions for State-constrained Optimal Control Problems . Journal of Optimization Theoryand Applications, 123 (3); 595–617, 2004.

c 2006 by M. Gerdts

Page 239: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 239/243

Page 240: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 240/243

232 BIBLIOGRAPHY

[Neu76] Neustadt, L. W. Optimization: A Theory of Necessary Conditions . Princeton, NewJersey, 1976.

[Obe86] Oberle, H. Numerical solution of minimax optimal control problems by multiple shooting technique . Journal of Optimization Theory and Applications, 50; 331–357,

1986.

[OG01] Oberle, H. J. and Grimm, W. BNDSCO – A Program for the Numerical Solution of Optimal Control Problems . Technical Report Reihe B, Bericht 36, HamburgerBeitrage zur Angewandten Mathematik, Department of Mathematics, University of Hamburg, http://www.math.uni-hamburg.de/home/oberle/software.html, 2001.

[PB93] Pacejka, H. and Bakker, E. The Magic Formula Tyre Model . Vehicle SystemDynamics, 21 supplement; 1–18, 1993.

[PB94] Pesch, H. J. and Bulirsch, R. The Maximum Principle, Bellman’s Equation and Caratheodorys Work . Journal of Optimization Theory and Applications, 80 (2);199–225, 1994.

[PBGM64] Pontryagin, L. S., Boltyanskij, V., Gamkrelidze, R. and Mishchenko, E. Mathema-tische Theorie optimaler Prozesse . Oldenbourg, Munchen, 1964.

[Pes78] Pesch, H. J. Numerische Berechnung optimaler Flugbahnkorrekturen in Echtzeit-Rechnung . Ph.D. thesis, Institut fur Mathematik, Technische Universitat Munchen,1978.

[Pes79] Pesch, H. J. Numerical computation of neighboring optimum feedback control schemes in real-time . Applied Mathematics and Optimization, 5; 231–252, 1979.

[Pes89a] Pesch, H. J. Real-time computation of feedback controls for constrained optimal control problems. I: Neighbouring extremals . Optimal Control Applications andMethods, 10 (2); 129–145, 1989.

[Pes89b] Pesch, H. J. Real-time computation of feedback controls for constrained optimal con-trol problems. II: A correction method based on multiple shooting . Optimal ControlApplications and Methods, 10 (2); 147–171, 1989.

[Pes02] Pesch, H. J. Schl¨ usseltechnologie Mathematik: Einblicke in aktuelle Anwendungen der Mathematik . B. G. Teubner, Stuttgart – Leipzig – Wiesbaden, 2002.

[Pet82a] Petzold, L. R. A description of DASSL: a differential/algebraic system solver . Rep.

Sand 82-8637, Sandia National Laboratory, Livermore, 1982.

[Pet82b] Petzold, L. R. Differential/algebraic equations are not ODE’s . SIAM Journal onScientific and Statistical Computing, 3 (3); 367–384, 1982.

[Pet89] Petzold, L. R. Recent developments in the numerical solution of differen-tial/algebraic systems . Computer Methods in Applied Mechanics and Engineering,75; 77–89, 1989.

[Pow78] Powell, M. J. D. A fast algorithm for nonlinearily constrained optimization calcu-lation . In Numerical Analysis (G. Watson, editor), volume 630 of Lecture Notes in Mathematics . Springer, Berlin-Heidelberg-New York, 1978.

c 2006 by M. Gerdts

Page 241: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 241/243

BIBLIOGRAPHY 233

[Pyt98] Pytlak, R. Runge-Kutta Based Procedure for the Optimal Control of Differential-Algebraic Equations . Journal of Optimization Theory and Applications, 97 (3);675–705, 1998.

[Qi93] Qi, L. Convergence analysis of some algorithms for solving nonsmooth equations .Mathematics of Operations Research, 18 (1); 227–244, 1993.

[QS93] Qi, L. and Sun, J. A nonsmooth version of Newton’s method . Mathematical Pro-gramming, 58 (3); 353–367, 1993.

[Ral94] Ralph, D. Global convergence of damped Newton’s method for nonsmooth equations via the path search . Mathematics of Operations Research, 19 (2); 352–389, 1994.

[Ris91] Risse, H.-J. Das Fahrverhalten bei normaler Fahrzeugf¨ uhrung , volume 160 of VDI Fortschrittberichte Reihe 12: Verkehrstechnik/Fahrzeugtechnik . VDI-Verlag, 1991.

[Rob76] Robinson, S. M. Stability Theory for Systems of Inequalities, Part II: Differentiable Nonlinear Systems . SIAM Journal on Numerical Analysis, 13 (4); 487–513, 1976.

[Roc70] Rockafellar, R. T. Convex Analysis . Princeton University Press, New Jersey, 1970.

[RS96] Ryoo, H. S. and Sahinidis, N. V. A branch-and-reduce approach to global optimiza-tion . Journal of Global Optimization, 8 (2); 107–138, 1996.

[RV02] Roubicek, T. and Valasek, M. Optimal control of causal differential-algebraic sys-tems . Journal of Mathematical Analysis and Applications, 269 (2); 616–641, 2002.

[Sah96] Sahinidis, N. V. BARON: A general purpose global optimization software package .

Journal of Global Optimization, 8 (2); 201–205, 1996.

[SB90] Stoer, J. and Bulirsch, R. Numerische Mathematik II . Springer, Berlin-Heidelberg-New York, 3rd edition, 1990.

[SBS98] Schulz, V. H., Bock, H. G. and Steinbach, M. C. Exploiting invariants in the numerical solution of multipoint boundary value problems for DAE . SIAM Journalon Scientific Computing, 19 (2); 440–467, 1998.

[Sch81] Schittkowski, K. The Nonlinear Programming Method of Wilson, Han, and Powell with an Augmented Lagrangean Type Line Search Function. Part 1: Convergence

Analysis, Part 2: An Efficient Implementation with Linear Least Squares Subprob-lems . Numerische Mathematik, 383; 83–114, 115–127, 1981.

[Sch83] Schittkowski, K. On the Convergence of a Sequential Quadratic Programming Method with an Augmented Lagrangean Line Search Function . Mathematische Op-erationsforschung und Statistik, Series Optimization, 14 (2); 197–216, 1983.

[Sch85] Schittkowski, K. NLPQL: A Fortran subroutine for solving constrained nonlinear programming problems . Annals of Operations Research, 5; 484–500, 1985.

[Sch94] Schittkowski, K. Parameter estimation in systems of nonlinear equations . Nu-merische Mathematik, 68; 129–142, 1994.

c 2006 by M. Gerdts

Page 242: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 242/243

Page 243: Optimal Control of ODES and DAES

8/19/2019 Optimal Control of ODES and DAES

http://slidepdf.com/reader/full/optimal-control-of-odes-and-daes 243/243

BIBLIOGRAPHY 235

[vS94] von Stryk, O. Numerische L¨ osung optimaler Steuerungsprobleme: Diskretisierung,Parameteroptimierung und Berechnung der adjungierten Variablen , volume 441 of VDI Fortschrittberichte Reihe 8: Meß-, Steuerungs- und Regeleungstechnik . VDI-Verlag, 1994.

[vS99] von Schwerin, R. Multibody System Simulation: Numerical Methods, Algorithms,and Software . volume 7 of Lecture Notes in Computational Science and Engineering .Springer, Berlin-Heidelberg-New York, 1999.

[Wer95] Werner, D. Funktionalanalysis . Springer, Berlin-Heidelberg-New York, 1995.