Home Laboratory Computer Lab Theses

3.5 Problems with equality and inequality constraints

We now turn our attention to problems of the calculus of variations in the presence of constraints. Similar to finite-dimensional optimization problems, it is more convenient to use Lagrange multipliers in order to derive the necessary conditions of optimality associated with such problems; these considerations are discussed in the following for equality constraints. The method of Lagrange multipliers is then used in to obtain necessary conditions of optimality for problems subject to (equality) terminal constraints and isoperimetric constraints, respectively. The Lagrange multiplier methodology is then extended to inequality constrained problems.

We start by considering a general class of equality constraints where the function space 𝒟 is defined by level-set curves of one ore more functionals G1,,Gng .

Definition 3.11: Equality constrained calculus of variations problem

Consider the problem to minimize:

minx(t)F[x]=t1t2f(t,x(t),x˙(t))dts.t.x𝒟𝒳

where 𝒟:={x𝒳such thatGi[x]=Kii=1,,ng} .

The functionals Gi are defined in 𝒳 as well.

Remark 3.19

Note that also the basic problem of Calculus of Variations we have addressed so far falls into this broader category. Consider the functionals G1[x]=x(t1) and G2[x]=x(t2) and recall that for the basic problem we had 𝒟:={x𝒞1[t1,t2]nxsuch thatx(t1)=x1,x(t2)=x2} . We can define the same 𝒟 as 𝒟:={x𝒞1[t1,t2]nxsuch thatG1[x]=x1andG2[x]=x2} .

Remark 3.20: Common types of constraint functionals

This kind of problems allow to treat more general problems such as endpoint constrained problems and isoperimetric problems. In the endpoint constrained case the constraint functionals Gi take the form Gi[x]=ψi(t2,x(t2)) for i=1,,nx . The constraint reduces to Gi[x]=0i=1,,nx , in this way the final state is forced to satisfy ψ(t2,x(t2))=0 . Isoperimetric problems are a class of problems where the constraint functionals are defined as G[x]=t1t2g(t,x,x˙)dt and the constraint is expressed as the level-set G[x]=L where L . These kind of constraints are found in minimum volume [area] problems with fixed area [perimeter] as for example the catenary or Dido’s problem that we will solve in the following sections.

The next theorem that we give without proof is instrumental to prove a fundamental lemma on the existence of minimizers of such equality constrained problems.

Theorem 3.11: Inverse Function Theorem

Let x0nx and η>0. If a function Φ:η(x0)nx has continuous first partial derivatives in each component with nonvanishing Jacobian determinant at x0, then Φ provides a continuously invertible mapping between η(x0) and a region containing a full neighborhood of Φ(x0) .

The following Lemma gives conditions under which a point x¯ in a normed linear space (𝒳;) cannot be a (local) extremal of a functional F , when constrained to the level set of another functional G .

Lemma 3.5

Let F and G be functionals defined in a neighborhood of x¯ in a normed linear space (𝒳;) , and let G[x]=K be the equality constraint defining 𝒟 . Suppose that there exist fixed directions ξ1,ξ2 such that the Gâteaux derivatives of F and G satisfy the Jacobian condition:

|δFx¯[ξ1]δFx¯[ξ2]δGx¯[ξ1]δGx¯[ξ2]|0

and are continuous in a neighborhood of x¯ (in each direction ξ1 , ξ2 ). Then, x¯ cannot be a local extremal for F when constrained to 𝒟:={x𝒳such thatG[x]=K} .

Proof. Consider the auxiliary functions j(η1,η2):=F[x¯+η1ξ1+η2ξ2] and g(η1,η2):=G[x¯+η1ξ1+η2ξ2] , we have j(0,0)=F[x¯] and g(0,0)=G[x¯]=K and by definition of Gâteaux derivative δFx¯[ξ1]=jη1|0,0=jη1(0,0) , δFx¯[ξ2]=jη2|0,0=jη2(0,0) , δGx¯[ξ1]=gη1|0,0=gη1(0,0) and δGx¯[ξ2]=gη2|0,0=gη2(0,0) . Let us also define the function Φ:22 as:

Φ=[j(η1,η2)g(η1,η2)],

the jacobian determinant of Φ at (0,0) is:

Det(JΦ(0,0))=|jη1(0,0)jη2(0,0)gη1(0,0)gη2(0,0)|=|δFx¯[ξ1]δFx¯[ξ2]δGx¯[ξ1]δGx¯[ξ2]|,

by hypothesis we have Det(J(0,0))0 and thence Theorem 3.11 is applicable. Therefore the pre-image points of the neighborhood of Φ(0,0)=[j(0,0),g(0,0)]=[F[x¯],K] maps a full neighbourhood of the origin (0,0) in the η1-η2 plane as shown in Figure 3.12. That means that we can find (η1×,η2×) and (η1o,η2o) such that:

j(η1o,η2o)<j(0,0)<j(η1×,η2×)g(η1o,η2o)=g(0,0)=g(η1×,η2×)=K,

by definition of the functions g and j we have:

F[x¯+η1oξ1+η2oξ2]<F[x¯]<F[x¯+η1xξ1+η2xξ2]G[x¯+η1oξ1+η2oξ2]<G[x¯]<G[x¯+η1xξ1+η2xξ2]=K,

that means that x~=x¯+η1oξ1+η2oξ2 is admissible (i.e. satisfy the constraint G[x~]=K ) and reduces the cost, that is F[x~]<F[x¯] hence x¯ cannot be a local extremal trajectory. □

pict

Figure 3.12:: Due to the inverse function theorem there exist a full neihborhood of (j(0,0),g(0,0)) hence it is possible to find trajectories with lower cost

With this preparation, it is easy to derive necessary conditions for a local extremal in the presence of equality or inequality constraints, and in particular the existence of the Lagrange multipliers.

Theorem 3.12: Existence of the Lagrange Multipliers

Let F and G be functionals defined in a neighborhood of x in a normed linear space (𝒳,), and having continuous Gâteaux derivatives in this neighborhood. Let also K=G(x) and suppose that x is a (local) extremal for F constrained to Γ(K):={x𝒳:G(x)=K} . Suppose further that δGx[ξ¯]0 for some direction ξ¯𝒳 . Then, there exists a scalar λ such that:

δFx[ξ]+λδGx[ξ]=0ξ𝒳aaNote that the variations need not to be admissible for the space Γ(K)..

Proof. Since x is a (local) extremal for F constrained to Γ(K) by Lemma 3.5 we must have that the determinant:

|δFx[ξ¯]δFx[ξ]δGx[ξ¯]δGx[ξ]|=0

for any ξ𝒳 Hence, by defining:

λ:=δFx[ξ¯]δGx[ξ¯]

it follows that δFx[ξ]+λδGx[ξ]=0 for each ξ𝒳 . □

Remark 3.21

As in the finite-dimensional case, the parameter λ in Theorem 3.12 is called a Lagrange multiplier. Using the terminology of directional derivatives appropriate to nx , the Lagrange condition δFx[ξ]=λδGx[ξ] says simply that the directional derivatives of F are proportional to those of G at x . Thus, in general, Lagrange’s condition means that the level sets of both F and G at x share the same tangent hyperplane at x i.e., they meet tangentially. Note also that the Lagrange’s condition can also be written in the form

δ(F+λG)x[ξ]=0ξ𝒳

This is due to the linearity of the Gâteaux derivative that is, given two scalars a1,a2 and two functionals A,B from 𝒳 to we have that δ(a1A+a2B)x[ξ]=a1δAx[ξ]+a2δAx[ξ] as shown in Remark 3.7 which suggests, as we will see in the following, consideration of the Lagrangian functional L:=F+λG .

It is possible, ableit technical, to extend the method of Lagrange multipliers to problems involving any finite number of constraint functionals:

Theorem 3.13: (Existence of the Lagrange Multipliers (Multiple Constraints)

Let F and Gi,i=1,ng be functionals defined in a neighborhood of x in a normed linear space (𝒳,) and having continuous Gâteaux derivatives in this neighborhood. Let also Ki=Gi[x],i=1,,ng and suppose that x is a (local) extremal for F constrained to Γ(K):={x𝒳:Gi[x]=Ki,i=1,,ng} . Suppose further that:

|δG1,x[ξ¯1]δG1,x[ξ¯ng]δGng,x[ξ¯1]δGng,x[ξ¯ng]|0

for ng (independent) directions ξ¯1,,ξ¯ng𝒳 . Then, there exists a constant vector λng such that:

δFx[ξ]+i=1ngλiδGi,x[ξ]=0ξ𝒳

Remark 3.22: Link to Nonlinear Optimization

The previous theorem is the generalization of Theorem 2.13 of Chapter 2 to optimization problems in normed linear spaces. Note, in particular, that the requirement that x to be a regular point for the Lagrange multipliers to exist is generalized by a non-singularity condition in terms of the Gâteaux derivatives of the constraint functionals G1,,Gng . Yet, this condition is in general not sufficient to guarantee uniqueness of the Lagrange multipliers.

Remark 3.23: Hybrid Method of Admissible Directions and Lagrange Multipliers

If x𝒟 with 𝒟 a subset of a normed linear space (𝒳,), and the 𝒟 -admissible directions form a linear subspace of 𝒳 (i.e., ξ1,ξ2𝒟η1ξ1+η2ξ2𝒟 for every scalars η1,η2 ), then the conclusions of Theorem 3.12 remain valid when further restricting the continuity requirement for F to 𝒟 and considering 𝒟 -admissible directions only. Said differently, Theorem 3.12 can be applied to determine (local) extremals to the functional F|𝒟 constrained to Γ(K) . This extension leads to a more efficient but admittedly hybrid approach to certain problems involving multiple constraints. Those constraints on which determine a domain 𝒟 having a linear subspace of 𝒟 -admissible directions, may be taken into account by simply restricting the set of admissible directions when applying the method of Lagrangian multipliers to the remaining constraints.

Necessary conditions of optimality for problems with end-point constraints and isoperimetric constraints shall be obtained with this hybrid approach in the next subsections.

3.5.1 Problems with end-point constraints

So far, we have only considered those problems with either free or fixed end-time t1 , t2 and end-points x(t1) , x(t2) . In this subsection, we shall consider problems having end-point constraints of the form ϕ(t2,x(t2))=0 with t2 being specified or not. In the case t2 is free, t2 shall be considered as an additional variable in the optimization problem. Like in 3.3.1 we shall then define the functions x(t) by extension on a ”sufficiently” large interval [t1,T] , and consider the linear space 𝒞1[t1,T]nx× , supplied with the weak norm (x,t)1,:=x1,+|t| . In particular, Theorem 3.13 applies readily by specializing the normed linear space (X,) to (𝒞1[t1,T]nx×,(,)1,) and considering the Gâteaux derivative δFx,t2[ξ,τ] at (x,t2) in the direction (ξ,τ) . These considerations yield necessary conditions of optimality for problems with end-point constraints as given in the following:

Theorem 3.14: Transversal Conditions

Consider the problem to minimize the functional

F[x,t2]:=t1t2f(t,x(t),x˙(t))dt

on 𝒟:={(x,t2)𝒞1[t1,T]nx×[t1,T]:x(t1)=x1,ϕ(t2,x(t2))=0} with f𝒞1([t1,T]×2×nx) and ϕ𝒞1([t1,T]×nx)nx . Suppose that (x,t2) gives a (local) minimum for F on 𝒟 and Rank[ϕtϕx]=nx at (x(t2),t2) . Then, x is a solution to the Euler equation of Theorem 3.3 satisfying both the end-point constraints x(t1)=x1 and the transversal condition:

[(ffx˙x˙)dt+fx˙dx]x,t2=0[dtdx]Ker[ϕtϕx]x,t2

in the particular one-dimensional case (i.e. nx=1 ) the transversal condition reduces to:

[ϕx(ff)ϕtf]x,t2=0.

Proof. Observe first that by fixing t2:=t2 and varying x in the 𝒟 -admissible direction ξ𝒞1[t1,t2]nx such that ξ(t1)=ξ(t2)=0 we show as in the proof of Theorem 3.3 that x must be a solution to the Euler equation on [t1,t2] . Observe also that the right end-point constraints may be expressed as the zero-level set of the functionals:

Gk[x]:=ϕk(t2,x(t2))=0k=1,,nx,

then simplifying the terms using the Euler equation as in the proof of Theorem 3.7 we obtain the first variations of the cost and constraint functionals as:

δF(x,t2)[ξ,τ]=f[x(t2)]x˙ξ(t2)+f[x(t2)]τδGk,(x,t2)[ξ,τ]=ϕk[x(t2)]tτ+ϕk[x(t2)]x(ξ(t2)+x˙(t2)τ)

where the usual compressed notation is used. Based on the differentiability assumptions on f and ϕ it is clear that these Gâteaux derivatives exist and are continuous. Further, since the rank condition Rank[ϕtϕx]=nx holds at (x(t2),t2) one can always find nx (independent) directions (ξ¯k,τ¯k)𝒞1[t1,t2]nx× such that the regularity condition:

|δG1,(x,t2)[ξ¯1,τ¯1]δG1,(x,t2)[ξ¯nx,τ¯nx]δGnx,(x,t2)[ξ¯1,τ¯1]δGnx,(x,t2)[ξ¯nx,τ¯nx]|0

is satisfied. Now, consider the linear subspace
Ξ:={(ξ,τ)𝒞1[t1,T]nx×:ξ(t1)=0} . Since (x,t2) gives a (local) minimum for F on 𝒟 by Theorem 3.13 and Remark 3.23 there exist a vector λnx such that:

0=δF(x,t2)[ξ,τ]+i=1ngλiδGi,(x,t2)[ξ,τ]==f[x(t2)]x˙ξ(t2)+f[x(t2)]τ++k=1nxλk(ϕk[x(t2)]tτ+ϕk[x(t2)]x(ξ(t2)+x˙(t2)τ)).

The above equation can be rewritten using the vectorial form as:

0=f[x(t2)]x˙ξ(t2)++(f[x(t2)]+λϕ[x(t2)]t)τ+λϕ[x(t2)]x(ξ(t2)+x2(t2)τ)

that must hold for every (ξ,τ)Ξ . In particular, we consider variations ξ such that ξ(t2)=x(t2)τ so that the second term is simplified and we get:

(f[x(t2)]f[x(t2)]x˙x˙(t2)+λϕ[x(t2)]t)τ=0

that must hold for each admissible τ and thus we get:

λϕ[x(t2)]t=f[x(t2)]+f[x(t2)]x˙x˙(t2)=bt.

Now considering variations such that τ=0 while ξi(t2) is such that ξi(t2)=(0,..,ξi(t2),..,0) we get for each i :

(f[x(t2)]i+λϕxi)ξi(t2)=0

and thus we have the equation:

λϕxi=f[x(t2)]i=bi,x

for each i=1,,nx . In vectorial notation we have the following nx dimensional system of equations:

λ[ϕtϕx]=[f[x(t2)]+f[x(t2)]x˙x˙(t2)f[x(t2)]x˙]

that following our definitions of bt and bx becomes:

λ[ϕtϕx]=[btbx]

that is a linear system of equations expressed as λA=[btbx] where A=[ϕtϕx] . Note that [ϕtϕx] is the jacobian Jϕ(t2,x(t2)) matrix whose dimensions are nx×nx+1 and by hypothesis its rank is nx . Therefore for the rank-nullity theorem dim(Ker(Jϕ(t2,x(t2))))=1 , we can therefore pick a nx+1 dimensional vector d=[dtdx]Ker(Jϕ(t2,x(t2))) and post-multiply by it the previous equation, noting that [ϕtϕx]d=0 since dKer(Jϕ(t2,x(t2))) , we get:

[btbx]d=0

note that if d belongs to the kernel of the Jacobian at (x(t2),t2) also d does, then it also holds:

[btbx]d=0

Substituting the definition of bt and bx we get:

[(ffx˙x˙)dt+fx˙dx]x,t2=0[dtdx]Ker[ϕtϕx]x,t2

Remark 3.24: one-dimensional case

For the one-dimensional case we have:

[(ff)dt+fdx]x,t2=0(dtdx)Ker[ϕtϕx]x,t2

note that [ϕtϕx]x,t2=v reduces to a row vector in 2 then the kernel is simply made of the subspace of vectors in d2 such that:

vd=0dtϕt+dxϕx=0,

by picking dx=ϕt and dt=ϕx the transversality condition reduces to:

[(ff)ϕxfϕt]x,t2=0

Remark 3.25

Note that ϕ𝒞1([t1,T]×nx)nx therefore its Jacobian is a matrix nx×(nx+1) of the form :

Jϕ=ϕ[tx]=[ϕ1tϕ1x1ϕ1xnxϕnxtϕnxx1ϕnxxnx]=[ϕtϕx]

and it is a matrix-valued function of (t,x) that is Jϕ(t,x) . Its rank can be at most nx when all of its rows are independent.

Example 3.15 (Minimum Path Problem with Variable End-Point and End-Time). Consider the problem to minimize the distance between a fixed pointA=(xA,yA) and the pointB=(xB,yB) {(x,y)2:y=ax+b} in the(x,y) -plane. We want to find the curvey(x) xAxxB such that the functionalF[y,xB]:=xAxB1+(x)2dx is minimized, subject to the constraintsy(xA)=yA andy(xB)=axB+b . Note that neitherxB nory(xB) are explicitly given. Instead they are implicitly given by the scalar end-point constraintϕ(xB,y(xB))=y(xB)axBb=0 whereϕ:2 is obviously a𝒞1 function. We stress the fact thatϕ(x,y)=0 (i.e. the zero level-set of the functionϕ ) is a straight line in2 . Therefore, we are looking for the minimum path problem from a fixed pointA to a pointB on a given straight line. For the sake of simplicity let’s considerA=(0,0) , from Theorem 3.14we know that an extremaly must satisfy the Euler equation in the intervalx[xAxB] . Note thatxB is yet unknown. As we have already seen the extremal for minimum path problems are straight lines. Since we have assigned a boundary condition inA , we get:

y(x)=C1x

that are a family of straight lines passing through the origin. In order to findC1 andxB we apply the necessary condition of Theorem 3.14adpated to the one dimensional case, that is:

[(ff)ϕyfϕx]y,xB=0

Substitutingf=1+C12 ,f=C11+C12 ,ϕy=1 ,ϕx=a we get:

1+C12C121+C12+aC11+C12=0

and upon simplification:

1+aC11+C12=0C1=1a

that precisely means that the extremal curvey=1ax is perpendicular to the straight line defined by the constraint. In this simple case, this is a global condition but the transversality condition in general is a local condition at the end-point between the ”gradient” of the constraint and the extremal trajectory. Note that we have also implicitly foundxB as the intersection between the linesy=1ax andy=ax+b that isxB=baa2+1 . A graphical representation of this problem is shown in Figure 3.13

pict

Figure 3.13:: Straight lines originating from the origin (i.e. point A ) in blue, optimal trajectory y in red, zero-level set of ϕ in green. Optimal point B=(y(xB),xB) as black dot. Simple case with a=1 and b=2 .

3.5.2 Problems with isoperimetric constraints

An isoperimetric problem of the calculus of variations is a problem wherein one or more constraints involves a functional in the form of an integral over part or all of the integration horizon [t1,t2] . Typically:

minx(t)𝒟F[x]=t1t2f(t,x(t),x˙(t))dt subject to: G[x]=t1t2ψ(t,x(t),x˙(t))dt=K

where K is a given number. Note that this problem already fits the framework that we have treated so far since the constraint is in the form of a level-set of the functional G[x] . We will use the hybrid approach in Remark 3.23 to deal both with simpler constraints imposed by the function space 𝒟 (e.g. fixed end-points) and with the isoperimetric constraint G[x]=K .

Theorem 3.15: First-Order Necessary Conditions for Isoperimetric Problems

Consider the problem to minimize the functional:

F[x]:=t1t2f(t,x(t),x˙(t))dt

on 𝒟:={x𝒞1[t1,t2]nx:x(t1)=x1,x(t2)=x2}, subject to the isoperimetric constraints:

Gi[x]:=t1t2ψi(t,x(t),x˙(t))dt=Ki,i=1,,ng

with f𝒞1([t1,t2]×2×nx) and ψi𝒞1([t1,t2]×2×nx),i=1,,ng. Suppose that x𝒟 gives a (local) minimum for this problem, and

|δG1,x[ξ¯1]δG1,x[ξ¯ng]δGng,x[ξ¯1]δGng,x[ξ¯ng]|0

for ng (independent) directions ξ¯1,,ξ¯ng𝒞1[t1,t2]nx . Then, there exists a constant vector λng such that x is a solution to the so called Euler-Lagrange’s equation:

ddti(t,x(t),x˙(t),λ)=xi(t,x(t),x˙(t),λ),i=1,,nx

where:

(t,x(t),x˙(t),λ):=f(t,x(t),x˙(t))+λψ(t,x(t),x˙(t))==f(t,x(t),x˙(t))+i=1ngλiψi(t,x(t),x˙(t)).

Proof. Remark first that, from the differentiability assumptions on f and ψi,i=1,,ng the Gâteaux derivatives δFx[ξ] and δGi,x[ξ] , i=1,,ng , exist and are continuous for every ξ𝒞1[t1,t2]nx .

Since x𝒟 gives a (local) minimum for F on 𝒟 constrained to Γ(K):={x𝒞1[t1,t2]nx:𝒢i(x)=Ki,i=1,,ng} , and x is a regular point for the constraints, by Theorem 3.13 (and Remark 3.23 ), there exists a constant vector λng such that:

δFx[ξ]+i=1ngλiδGi,x[ξ]=0

for each 𝒟 -admissible direction ξ . Observe that this latter condition is equivalent to that of finding a minimizer to the functional:

F[x]:=t1t2f(t,x(t),x˙(t))+λψ(t,x(t),x˙(t))dt:=t1t2(t,x(t),x˙(t),λ)dt

on 𝒟 . The conclusion then directly follows upon applying Theorem 3.3. □

Remark 3.26: First Integrals

Similar to free problems of the calculus of variations (see Remark 3.11) it is easy to show that the Hamiltonian function H(x(t),x˙(t),λ) defined as

H:=x˙x˙

is constant along an extremal trajectory provided that does not depend on the independent variable t . Recall the definition =f(x(t),x˙(t))+λψ(x(t),x˙(t)) and that along an extremal trajectory the necessary condition of Theorem 3.15 holds, that is:

ddtx˙x=0.

The total time derivative of the Hamiltonian is:

dHdt=xx˙+x˙x¨ddtx˙x˙x˙x¨=0

Example 3.16 (Solution to Dido’s Problem). We are finally ready to face problem shown in Example 3.3. Dido’s Problem is to minimize the functional:

maxy𝒟F[y]=xAxBydx8

subject to the isoperimetric 9 constraint:

G[y]=xAxBds=xAxB1+2dx=L

The subspace𝒟 is𝒟={y𝒞1[xA,xB]such thaty(xA)=yAy(xB)=yB} . Now we build the function as:

=y+λ1+2

Since is independent ofx , the hamiltonian function is constant along an extremal trajectory, that is:

H==C1

in this case we have=λ1+2 , hence:

y+λ1+2λ21+2=C1.

Manipulating the previous relation we get:

(y+λ1+2)1+2λ2=C11+2(C1y)1+2=λ

that is a separable nonlinear differential equation, upon squaring both terms and after some algebraic manipulation we get:

=λ2(C1y)2(C1y)2

and by formally expressing=dydx and integrating both sides we have:

yAyC1yλ2(C1y)2dy=xBxdx.

In order to solve the integral indy we make the substitutiony=C1λcos(𝜃) ,dy=λsin(𝜃)d𝜃 and hence:

𝜃A𝜃λcos(𝜃)λ2λ2cos2(𝜃)λsin(𝜃)d𝜃=𝜃A𝜃λcos(𝜃)d𝜃

in a similar way as for the brachistochrone problem we get the extremal trajectory in parametric form as:

y(𝜃)=C1+λ^cos(𝜃)x(𝜃)=C2+lambda^sin(𝜃)

where we have definedλ^=λ . Note that(x,y) define a circular arc of radiusλ^ and centered at(C2,C1) . This is easy to see considering that:

yC1=λ^cos(𝜃)xC2=λ^sin(𝜃)

and thus:

(yC1)2+(xC2)2=λ^2

however we still need to compute the actual value of the radiusλ^ and the center coordinates(C2,C1) . Assume for the sake of simplicity thatxA=1 ,xB=1 andyA=yB=0 . Substituing the boundary condtions in the previous equation gives:

(C1)2+(1C2)2=λ^2(C1)2+(1C2)2=λ^2

by subtracting these two conditions we have:

(1C2)2(1C2)22C2=0C2=0

therefore the center of this family of circles is in(0,C1) and passes throughA=(1,0) andB=(1,0) . We still need to compute the radiusλ^ and the ordinate of the centerC1 . Indeed these two parameters depend on the isoperimetric constraint (i.e. the length of the circle)L that is given. The length of an arc of circle of radiusλ^ between𝜃A and𝜃B (yet unknown) isL=λ^(𝜃B𝜃A) , this comes from the definition of radiants, obviously the same relation holds if we compute it analytically as:

G[y(𝜃),x(𝜃)]=𝜃A𝜃Bx𝜃2+y𝜃2d𝜃=λ(𝜃B𝜃A)=L

𝜃A and𝜃B can expressed in terms ofC1 by using the boundary condition atB andA :

x(𝜃B)y(𝜃B)C1=1C1=λ^sin(𝜃B)λ^cos(𝜃B)=tan(𝜃B)x(𝜃A)y(𝜃A)C1=1C1=λ^sin(𝜃A)λ^cos(𝜃A)=tan(𝜃A).

Since the arctangent function is an odd function we have𝜃A=𝜃B=arctan(1C1) while from the cartesian expression of the circle withC2 = 0 at pointB we get:

C12+1=λ^2λ^=C12+1

finally the expression for the lengthL presents the only unknownC1 :

L=2C12+1arctan(1C1)

that, givenL , can be solved numerically forC1 that is the ordinate of the center of the circle, then usingλ=C12+1 we can retrieve the value of the Lagrange multiplierλ that, interestingly enough, in this case is also the radius of the circle changed of sign.L as a function ofC1 is plotted in Figure 3.14. The extremal for the simple caseC1=1 that givesλ=2 and𝜃B=π4 is plotted in Figure 3.15.

pict

Figure 3.14:: Plot of the function y=2C12+1arctan(1C1) , note that we are not completely free in choosing the arc length L since as C10 Lπ while C1 L2 that is 2<L<π for the problem to be well-posed.

pict

Figure 3.15:: Extremal trajectory in Red from point A to B, geometric construction in Blue.

Example 3.17 (Problem of the Surface of Revolution of Minimum Area). Consider the problem to find the smooth curvey(x)0 having a fixed lengthL>0, joining two given pointsA=(xA,yA) andB=(xB,yB), and generating a surface of revolution around thex -axis of minimum area. In mathematical terms, the problem consists of finding a minimizer of the functional:

F[y]:=2πxAxBy(x)1+(x)2dx

on𝒟:={x𝒞1[xA,xB]:y(xA)=yA,y(xB)=yB}, subject to the isoperimetric: constraint

G[y]:=xAxB1+(x)2dx=μ.

Let us drop the coefficient2π inF and introduce the Lagrangian as

(y(x),(x),λ):=y(x)1+(x)2+λ1+(x)2

since does not depend on the independent variablex we use the fact thatH must be constant along an extremal trajectoryȳ(x) ,

H:==C1

for some constantC1. That is:

(ȳ(x)+λ)1+ȳ˙(x)2(ȳ(x)+λ)ȳ˙21+ȳ˙(x)2=C1

that can be rearranged as:

(ȳ(x)+λ)(1+ȳ˙(x)2)(ȳ(x)+λ)ȳ˙21+ȳ˙(x)2=(ȳ(x)+λ)1+ȳ˙(x)2=C1

that with some simplification gives:

(x)C12(λ+y)2C12=dx

again we solve the integral by substitutingy=C1cosh(𝜃)λ anddy=C1sinh(𝜃)d𝜃 and thus:

yAyC12(λ+y)2C12dy=𝜃A𝜃C12(C1cosh(𝜃)2)2C12C1sinh(𝜃)d𝜃

and by using the relationcosh(𝜃)2sinh(𝜃)2=1 we obtain:

𝜃A𝜃C12(C1cosh(𝜃)2)2C12C1sinh(𝜃)d𝜃=C1𝜃+C2

and assumingxA=0 we have:

x=C1𝜃+C2𝜃=xC2C1

finally substituing back iny=C1cosh(𝜃)λ we obtain:

ȳ(x)=C1cosh(xC2C1)λ.

The constantsC1 ,C2 andλ are to be found from the boundary conditions and the isoperimetric constraints. Note that the extremal curve are a family of catenaries. For the sake of simplicity let’s do the the inverse reasoning that is we assignC1,C2,λ and we compute the boundary condititions. Note thatC2 is a translation along thex -axis whileλ a translation along they -axis.WithC1=C2=10 ,λ=5.4 andxA=0 ,xB=20 we get the results of Figure 3.16and 3.17where the lengthL of the curve is:

L=G[ȳ]=xAxB1+sinh(xC2C1)2dx=23.5

note that the lengthL does not depend directly on the lagrange multiplierλ . While the surface area is:

A=F[ȳ]=2πxAxBȳ(x)1+ȳ˙(x)2dx=970.26

pict

Figure 3.16:: Generating curve ȳ

pict

Figure 3.17:: Generated surface of revolution of minimum area

Example 3.18 (Solution to hanging rope with fixed length). We give the solution to the problem shown in Example 3.2. Given a rope of lengthL attached at two polesA=(xA,yA) andB=(xB,yB) find the functiony(x) that describes the shape assumed by the rope under the action of gravity. We want to minimize the functional expressing the potential energy under the isoperimetric constraint expressing the fixed length. That is:

F[y]:=xAxBy(x)1+(x)2dx

on𝒟:={x𝒞1[xA,xB]:y(xA)=yA,y(xB)=yB}, subject to the isoperimetric constraint

G[y]:=xAxB1+(x)2dx=μ

This is exactly the same problem as the surface of minimum area. The rope assumes the shape of a catenary:

y(x)=C1cosh(xC2C1)λ

where againC1,C2,λ are to be found from the boundary condition and the lengthL .

3.5.3 Problems with inequality constraints

The method of Lagrange multipliers can also be used to address problems of the calculus of variations having inequality constraints (or mixed equality and inequality constraints), as shown by the following:

Theorem 3.16: Existence and Uniqueness of the Lagrange Multipliers (Inequality Constraints)

Let F and G1,,Gng be functionals defined in a neighborhood of x in a normed linear space (𝒳,), and having continuous and linear Gâteaux derivatives in this neighborhood. Suppose that x is a (local) minimum point for F constrained to Γ(K):={x𝒳:𝒢i(x)Ki,i=1,,ng} for some constant vector K . Suppose further that nang constraints, say G1,,Gna for simplicity, are active at x and satisfy the regularity condition

|δG1,x[ξ¯1]δG1,x[ξ¯na]δGna,x[ξ¯1]δGna,x[ξ¯na]|0

for na (independent) directions ξ¯1,,ξ¯na𝒳 (the remaining constraints being inactive). Then, there exists a vector νng such that:

δFx[ξ¯]+i=1ngδGi,x[ξ¯]νi=0ξ𝒳

and

νi0(Gi[x]Ki)νi=0

for i=1,,ng

Proof. Since the inequality constraints Gna+1,,Gng are inactive, the nonnegativity conditions and complementary slackness are trivially satisfied by taking νna+1==νng=0 . On the other hand, since the inequality constraints G1,,Gna are active and satisfy a regularity condition at x , the conclusion that there exists a unique vector λna such that the ”stationarity condition” holds follows from Theorem 3.12 moreover, the complementarity slackness condition is trivially satisfied for i=1,na since Gi[x]=Ki if i is active. Hence, it suffices to prove that the Lagrange multipliers νiνna cannot assume negative values when x is a (local) minimum.

We show the result by contradiction. Without loss of generality, suppose that ν1<0 , and consider the (na+1)×na matrix A defined by

A=[δFx[ξ¯1]δFx[ξ¯na]δG1,x[ξ¯1]δG1,x[ξ¯na]δGna,x[ξ¯1]δGna,x[ξ¯na]].

By hypothesis, Rank(A)na1 , hence the null space of A has dimension lower than or equal to 1 . But from the stationarity condition the nonzero vector (1,ν1,,νna)Ker(A) . That is Ker(A) has dimension equal to 1 and Ay<0 only if η such that y=η(1,ν1,,νna) . Because ν1<0 there does not exist a y0 in Ker(A) such that yi0 for every i=1,,na+1 . Thus, by Gordan’s Theorem there exists a nonzero vector p0 in na such that Ap<0 , or equivalently:

k=1napkδFx[ξ¯k]<0k=1napkδGi,x[ξ¯k]<0i=1,,na.

The Gâteaux derivatives of F,G1,,Gna being linear (by assumption), we get:

δFx[k=1napkξ¯k]<0δGi,x[k=1napkξ¯k]<0,i=1,,na.

In particular,

δ>0 such that x+ηk=1napkξ¯kΓ(K),0ηδ

that is, x being a local minimum of F on Γ(K) ,

δ(0,δ] such that F[x+ηk=1napkξ¯k]F[x]0ηδ

and we get

δFx[k=1napkξ¯k]=limη0+F[x+ηk=1napkξ¯k]F[x]η0

which contradicts the inequality obtained earlier. □