the solution of euclidean norm trust region sqp ...the presentation is in- ... november 11, 2016 1....

25
The Solution of Euclidean Norm Trust Region SQP Subproblems via Second Order Cone Programs, an Overview and Elementary Introduction Florian Jarre, Felix Lieder, Mathematisches Institut, Heinrich-Heine Universit¨ at D¨ usseldorf, Germany. Abstract It is well known that convex sequential quadratic programming (SQP) subproblems with a Euclidean norm trust region constraint can be reduced to second order cone programs for which the theory of Euclidean Jordan algebras leads to efficient interior-point algo- rithms. Here, a brief and self-contained outline of the principles of such an implementation is given. All identities relevant for the im- plementation are derived from scratch and are compared to interior- point methods for linear programs. Sparsity of the data of the SQP subproblem can be maintained essentially in the same way as for interior-point methods for linear programs. The presentation is in- tended as an introduction for students and for colleagues who may have heard about Jordan algebras but did not yet find the time to get involved with them. A simple Matlab implementation is made available and the discussion of implementational aspects addresses a scaling property that is critical for SQP subproblems. Key words: Second order cone program, Jordan algebra, SQP subproblem. November 11, 2016 1

Upload: others

Post on 14-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

The Solution of Euclidean Norm Trust Region

SQP Subproblems via Second Order Cone Programs,

an Overview and Elementary Introduction

Florian Jarre, Felix Lieder,Mathematisches Institut,Heinrich-Heine Universitat Dusseldorf, Germany.

Abstract

It is well known that convex sequential quadratic programming(SQP) subproblems with a Euclidean norm trust region constraintcan be reduced to second order cone programs for which the theoryof Euclidean Jordan algebras leads to efficient interior-point algo-rithms. Here, a brief and self-contained outline of the principles ofsuch an implementation is given. All identities relevant for the im-plementation are derived from scratch and are compared to interior-point methods for linear programs. Sparsity of the data of the SQPsubproblem can be maintained essentially in the same way as forinterior-point methods for linear programs. The presentation is in-tended as an introduction for students and for colleagues who mayhave heard about Jordan algebras but did not yet find the time toget involved with them. A simple Matlab implementation is madeavailable and the discussion of implementational aspects addressesa scaling property that is critical for SQP subproblems.

Key words: Second order cone program, Jordan algebra, SQP subproblem.

November 11, 2016

1

Page 2: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

1 Introduction

The trust region subproblem for unconstrained minimization typically is statedin terms of minimizing a quadratic function subject to a Euclidean norm trustregion constraint, see e.g. [9]. This trust region problem is “reasonably easy” tosolve (see e.g. [25]) while minimizing a quadratic function subject to an∞-normtrust region is NP -complete. Moreover, if the trust region constraint is active,then the solutions of trust region subproblems with a polyhedral trust regiongenerally lie at a vertex or a low dimensional facet of the trust region, and thereis no compelling reason why one would want to restrict the search directions tosuch low dimensional facets that depend on the particular choice of the trustregion. Under mild conditions, the Euclidean norm trust region subproblemresults in globally convergent and rapidly locally convergent algorithms.

The transfer of Euclidean norm trust region subproblems to equality con-strained problems has been considered in a series of papers [7, 33, 42, 45, 8, 31,35, 4].

For minimization subject to equality and inequality constraints the situa-tion often is perceived differently: As examples by Meredith Goldsmith [17]illustrate, a nonconvex SQP trust region subproblem may generate search di-rections that (locally and globally) neither reduce the objective function nor theunderlying measure for the constraint violation. One way to cope with this sit-uation is to define convex sequential quadratic programming (SQP) trust regionsubproblems in a way that local superlinear convergence can still be anticipated.For such trust region subproblems, often a 1-norm or an ∞-norm trust regionconstraint is considered, e.g. [9, 14, 43]. This choice allows that the trust re-gion problem can be converted to an equivalent convex QP which is also easilysolvable. As far as the global convergence analysis is concerned, since all normsare equivalent, the choice of the norm is not crucial. Likewise, fast local con-vergence is expected only when the trust region constraint is not active in thefinal iterations, and thus, the choice of the norm is not crucial either. Never-theless it is somewhat unsatisfactory to prefer the Euclidean norm trust regiononly in the case of unconstrained (or equality constrained) minimization. Here,the known fact is repeated that also in the presence of inequality constraints,a Euclidean norm trust region subproblem is easy to solve, not only in theory,but also in practice. To support the latter point, a self-contained derivation ofan interior-point approach for second order cone programs is presented and itsapplication to a Euclidean norm trust region subproblem.

The papers [33, 43] each present an appealing global convergence analysisfor a p-norm SQP method. As an alternative to such methods several otherSQP approaches that do not rely on lp-norm trust region subproblems have beenproposed such as constraint relaxation [39], constraint reduction [30], constraintlumping [7], or the filter-SQP approach [15]. For a detailed discussion we refer,for example, to the monograph [18].

Optimization algorithms over the second order cone (SOC) are well studied,see e.g. [12, 13, 28, 29, 36, 24, 34, 38, 5], and the underlying monograph [11]. Acomprehensive survey with numerous further applications of second order coneprograms is given in [3], see also [22]. In this paper, based on [19, 10], abrief and self-contained outline of the principles of an implementation is given.The results needed for the design and analysis of an interior-point algorithmare derived from scratch and in analogy to interior-point methods for linear

2

Page 3: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

programs (LPs). As an application, the solution of penalty SQP subproblemswith a Euclidean trust region constraint is discussed.

1.1 Notation

The components of a vector x ∈ Rn are denoted by xi for 1 ≤ i ≤ n. TheEuclidean norm of a vector x ∈ Rn is denoted by ‖x‖ and Diag(x) denotes then×n diagonal matrix with diagonal entries xi for 1 ≤ i ≤ n. The identity matrixis denoted by I, its dimension being evident from the context. Inequalities suchas x ≥ 0 are understood componentwise. The second order cone is given by

SOC =

{x ∈ R1+n | x =

(x0

xN

), x0 ∈ R, xN ∈ Rn, ‖xN‖ ≤ x0

}. (1)

Here, N := {1, . . . , n} spans the indices from 1 to n. The partition of a vectorx ∈ R1+n into a first component x0 ∈ R and a vector xN ∈ Rn will be usedthroughout sections 2 and 3 of this paper. The interior of SOC is denoted bySOC◦.

2 Second order cone programs

A second order cone program (SOCP) is the problem of minimizing a linearobjective function over the Cartesian product of second order cones and subjectto linear equality constraints. In the following, the theory of second order coneprograms is contrasted with linear programs and in particular with interior-pointmethods for solving linear programs.

A crucial feature in the design of long step interior-point methods for linearprograms is the fact that the positive orthant Rn

+ is selfdual, i.e. x ∈ Rn+ if, and

only if, xT s ≥ 0 for all s ∈ Rn+. This implies that the primal and dual conic

constraints for linear programs take the same form, namely x ≥ 0 and s ≥ 0where x and s are the primal and dual variables. As a direct consequence of theCauchy-Schwarz inequality, also SOC is selfdual.

For the comparison with linear programs in the next section, problems in-volving a single second order cone of the form (1) are considered; the general-ization to the Cartesian product of several second order cones is addressed inSection 3.3.

2.1 Relation to linear programs

First, using a notation that is suitable also for second order cone programs,a brief summary of some well known principles of interior-point methods forsolving linear programs is presented.

Let A ∈ Rm×n, b ∈ Rm, c ∈ Rn be given. For a linear program in standardform,

(LP ) min { cTx | Ax = b, x ≥ 0 },

a point x ≥ 0 with Ax = b is optimal, if, and only if, there exists a dual variables ≥ 0 that satisfies AT y + s = c for some y, as well as the complementaritycondition xT s = 0.

3

Page 4: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

In this subsection let “◦” denote the Hadamard-product (the componentwiseproduct) of two vectors. Then, for a vector x > 0, powers of x (with respectto the multiplication “◦”) can be defined in an obvious fashion, for examplex1/2 ≥ 0 with x1/2 ◦ x1/2 = x. Evidently,

Rn+ := {x ∈ Rn | x ≥ 0} = {v ◦ v | v ∈ Rn} (2)

showing that the positive orthant is the cone of squares with respect to the mul-tiplication “◦”. In the next subsection another product “◦” will be consideredleading to SOC as a cone of squares. Interior point methods for linear programsuse the following reformulation of the complementarity condition,

x ≥ 0, s ≥ 0, xT s = 0 ⇐⇒ x ≥ 0, s ≥ 0, x ◦ s = 0. (3)

The right hand side converts the scalar complementarity condition to n bilin-ear equations. These are perturbed and linearized at each step of primal-dualinterior-point algorithms. More precisely, at each iteration of an interior-pointalgorithm, given some iterate x, y, s with x, s > 0, systems of the following formfor finding a correction ∆x,∆y,∆s are considered:

A∆x = b−AxAT ∆y + ∆s = c−AT y − s (4)

s ◦∆x+ x ◦∆s = µe− x ◦ s

where, in this subsection, e = (1, . . . , 1)T is the neutral element with respect tothe multiplication “◦”. Changing the multiplication “◦” in the next subsectionalso results in a change of the neutral element e.

Operations such as s ◦∆x are typically expressed by a multiplication with adiagonal matrix, s◦∆x ≡ Diag(s)∆x; the above presentation is used to contrastthe situation with the second order cone.

A crucial element in solving (4) is the fact that the last line can easily besolved for ∆x (or for ∆s):

∆x = s−1 ◦ (s ◦∆x) = s−1 ◦ (µe− x ◦ s− x ◦∆s), (5)

where s−1 is the componentwise inverse of s satisfying s−1◦s = e. Note that thefirst equation in (5) exploits the fact that the Hadamard-product is associative.Solving the second line for ∆s and inserting both into the first row leads to alinear system of the form

AD2AT ∆y = rhs (6)

where D = DT = Diag(x1/2 ◦ s−1/2) is a scaling matrix satisfying the equation

Ds = D−1x, (7)

see e.g. [40].

2.2 Transfer to second order cone programs

We now turn from (LP ) to second order cone programs where the self-dual coneRn

+ := {x ∈ Rn | x ≥ 0} of the linear program is replaced with the second ordercone SOC, leading to

(SOCP ) min { cTx | Ax = b, x ∈ SOC }.

4

Page 5: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

Here, A ∈ R(1+n)×m, b ∈ Rm, and c ∈ R1+n. SOC being self-dual, it turnsout that the optimality conditions1 can be written in the same fashion as for(LP ), namely, if both, primal and dual, have a strictly feasible solution, thenx ∈ SOC with Ax = b is optimal if, and only if, there exist s ∈ SOC and ysatisfying AT y + s = c, as well as the complementarity condition xT s = 0, seee.g. [27], Theorem 4.2.1.

Interior-point methods for linear programs use the representation (2) of Rn+

as a cone of squares leading to the scaling matrix D in (7). It turns out that(2) and (7) can be transferred to SOCPs (and to semidefinite programs) whenchanging the product “◦”. For x, s ∈ R1+n partitioned as in (1), let the Jordan-product of x and s be given by

x ◦ s :=

(xT s

x0sN + s0xN

), and let L(x) :=

x0 x1 · · · xnx1 x0 0 0... 0

. . . 0xn 0 0 x0

(8)

so that x ◦ s = L(x)s for all x, s. (For linear programs, the matrix L(x) isreplaced with the matrix Diag(x).) Note that the Jordan-product “◦” is bi-linear and commutative (x ◦ s = s ◦ x), but – unlike the Hadamard-productand unlike the standard matrix multiplication – it is not associative (in general,x ◦ (y ◦ z) 6= (x ◦ y) ◦ z). Let further xN := (0, xTN )T and

e := (1, 0, . . . , 0)T .

Then,L(x) = x0I + exTN + xNe

T (9)

from which it is easy to derive that L(x) has the eigenvalues x0 and x0 ±‖xN‖. The Jordan-product replaces the Hadamard-product used in interiorpoint methods for linear programs. In fact, it is easy to verify that in analogyto (2)

SOC = {v ◦ v | v ∈ R1+n}.

And, as in (3), the complementarity conditions can be rewritten as

x ∈ SOC, s ∈ SOC, x ◦ s = 0.

However, due to the lack of associativity of the Jordan-product, the first equa-tion in (5) no longer holds true for the Jordan-product. The main difference inthe design and analysis of interior-point methods for SOC problems comparedto linear programs is to compensate for this lack of associativity.

The most evident way to compensate might seem to use the relation s◦∆s ≡L(s)∆s and to replace (5) with

∆x = L(s)−1(L(s)∆x) = L(s)−1 ◦ (µe− x ◦ s− x ◦∆s). (10)

1Note that (SOCP ) in the form stated here can be solved directly; the techniques outlinedbelow, however, can be extended in a straightforward way to problems over the Cartesianproduct of several second order cones. A second order cone of dimension 1 (i.e. where n = 0in (1)) is just a non-negative variable, so that the theory covers mixed linear - second ordercone programs, referred to as (MSOCP ) below, as well.

5

Page 6: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

It will turn out, however, that there is a more natural way to transfer relation(5) to SOCPs.

Note that analogously to e in (4), the vector e = (1, 0, . . . , 0)T used in (9)is the neutral element of the Jordan-product, i.e. L(e) = I the identity matrix.By definition, e2 = e, so that e lies in SOC, the cone of squares. In addition, eis also a “central element” in SOC, maximizing the Euclidean distance to theboundary of SOC within the set {x ∈ SOC | ‖x‖ ≤ 1}. This, and the fact thatthe Jordan product is “compatible” with the scalar product associated with theEuclidean norm, in the sense that xT (y ◦ z) = (x ◦ y)T z for all x, y, z ∈ R1+n, isessential for a theoretical analysis (and the practical efficiency) of the algorithmin Section 3.

2.3 Computations with the Jordan-product

It is also straightforward to transfer the notion of the inverse and the squareroot from the Hadamard-product to the Jordan-product: First, it is easy toverify that

x ◦ x− 2x0︸︷︷︸=:trace(x)

x+ (x20 − xTNxN )︸ ︷︷ ︸=:det(x)

e = 0 (11)

for all x ∈ R1+n. Relation (11) defines the trace and determinant for the Jordan-product2, in particular,

det(x) := x20 − xTNxN

for any x ∈ R1+n.From (9) and the Sherman-Morrison-Woodbury update formula one obtains

L(x)−1 =1

x0

(I − 1

det(x)

(x0(exTN + xNe

T )− xN xTN − ‖xN‖2eeT))

(12)

=1

x0

((0

I

)+

1

det(x)

(−x0

xN

)(−x0

xN

)T)

for x with x0 6= 0 and det(x) 6= 0, while the inverse element with respect to theJordan-product of a vector x with det(x) 6= 0 is defined by the characteristicequation x−1 ◦ x = e. This characteristic equation implies

x−1 = L(x−1)e =1

det(x)

(x0

−xN

).

and if x0 6= 0, also the relation x−1 = L(x)−1e holds true. Note that in general,L(x−1) 6= L(x)−1 and det(x) 6= det(L(x)). (The former is due to the lack ofassociativity of the Jordan-product.) Not only the inverse, but more generally,any analytic function R→ R can be extended to vectors in R1+n based on theeigenvalue decomposition of a vector x with xN 6= 0,

x =

(x0

xN

)= (x0 + ‖xN‖)

(12xN

2‖xN‖

)︸ ︷︷ ︸

=:u(1)

+(x0 − ‖xN‖)

(12−xN

2‖xN‖

)︸ ︷︷ ︸

=:u(2)

. (13)

2Relation (11) is called the characteristic polynomial of x transferring the notion of char-acteristic polynomial from the matrix multiplication to the Jordan-product. For the Jordan-product related to SOC the characteristic polynomial always has degree two.

6

Page 7: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

Here, u(1), u(2) ∈ SOC with u(1) + u(2) = e, u(1) ◦ u(1) = u(1), u(2) ◦ u(2) = u(2),and u(1) ◦ u(2) = 0 are the eigenvectors3 of x and x0 ± ‖xN‖ are the associatedeigenvalues, the product of which is the determinant, and the sum of which isthe trace. Let

ξ :=√x0 + ‖xN‖+

√x0 − ‖xN‖ =

√2x0 + 2

√x2

0 − ‖xN‖2.

The orthonormality of the eigenvectors u(i) with respect to the Jordan-productimplies that for x ∈ SOC, the root

x1/2 :=√x0 + ‖xN‖u(1) +

√x0 − ‖xN‖u(2) =

(ξ/2

xN/ξ

)∈ SOC (14)

satisfies x1/2 ◦ x1/2 = x. (The equation v ◦ v = x generally has four solutions;the one in (14) is called root of x.) In the same manner as the square root,any other analytic function R→ R can be applied to the eigenvalues of x, thusdefining an analytic function R1+n → R1+n.

We stress that operations such as x1/2 depend on the Jordan-product “◦”;the square root of the Hadamard-product (for linear programs) is not to beconfused with the square root in (14).

2.4 Quadratic representation

In the following, the derivative of the map f inv : R1+n\{x | det(x) = 0} → R1+n

with x 7→ x−1 will play an important role. f inv being a rational function of thevector entries, it is differentiable on its domain and its derivative satisfies theidentity

e+ o(‖t∆x‖) ≡ (x+ t∆x) ◦ (x−1 +Df inv(x)[t∆x])︸ ︷︷ ︸=(x+t∆x)−1+o(‖t∆x‖)

≡ e+ t∆x ◦ x−1 + x ◦Df inv(x)[t∆x] + o(‖t∆x‖)

for all x with det(x) 6= 0 and all ∆x and all t sufficiently close to zero. Sub-tracting e from both sides, dividing by t, and taking the limit as t→ 0 yields

0 = ∆x ◦ x−1 + x ◦Df inv(x)[∆x].

This equation – being satisfied for any ∆x – defines the linear map Df inv(x).Due to the lack of associativity this equation cannot be solved for Df inv(x)[∆x]by multiplying from left with x−1. Instead, denote the inverse4 of −Df inv(x)by P (x) := −(Df inv(x))−1 and y := Df inv(x)[∆x]. Then, above equation canbe written as a characteristic equation for P ,

(P (x)y) ◦ x−1 = x ◦ y. (15)

3For semidefinite programs, the Jordan-product of two symmetric matrices X,S is X ◦S :=12

(XS + SX) where XS is the usual matrix multiplication. If some vectors u(i) form a basis

of orthonormal eigenvectors of X, then the “eigenmatrices” of X are given by u(i)(u(i))T

satisfying the relation X ◦u(i)(u(i))T = λiu(i)(u(i))T . This latter relation corresponds to the

above definition of eigenvectors associated with the Jordan-product in (8).4Here, the inverse refers to the linear operator −Df inv(x). In general the form of the

inverse will be evident from the context; an inverse of a vector referring to the inverse withrespect to the Jordan-product, and an inverse of a matrix referring to the inverse of a linearmapping.

7

Page 8: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

It turns out that

P (x) = 2L(x)2 − L(x2) = det(x)(I − 2eeT ) + 2xxT (16)

satisfies this equation: Indeed, the second equality in (16) follows from (9), andit leads to

P (x)y = 2xT yx+ det(x)

(−y0

yN

)(17)

and

(P (x)y) ◦ x−1 = 2xT y

det(x)

(x0

xN

)◦(x0

−xN

)+

(−y0

yN

)◦(x0

−xN

)= x ◦ y

for any x, y with det(x) 6= 0 so that also the characteristic equation (15) holdstrue. 2

Here, P is “some form of square” of x satisfying the equation P (x)x−1 = xand is called quadratic representation of the Jordan algebra associated with theJordan-product “◦”. A key relation is the property

(P (u)v)−1 = P (u−1)v−1 (18)

for u, v with det(u) 6= 0 6= det(v).To prove (18) observe that the equation (18) is invariant under multiplication

of u or v with a nonzero constant. Thus, for the proof, one can assume withoutloss of generality that det(u) = det(v) = 1. In this case, the relation uT v =(u−1)T v−1 holds and thus, using (17), one obtains

(P (u)v) ◦ (P (u−1)v−1) = (2uT vu− v−1) ◦ (2uT vu−1 − v)

=

(2uT vu0 − v0

2uT vuN + vN

)◦(

2uT vu0 − v0

−2uT vuN − vN

)= e. 2

Finally, using the Sherman-Morrison rank-one update formula for inverse ma-trices one can verify that

P (u−1) = P (u)−1 for det(u) 6= 0, and P (u1/2) = P (u)1/2 for u ∈ SOC◦,(19)

where SOC◦ denotes the interior of SOC.

2.5 Scaling point

For x, s ∈ SOC◦, the interior of SOC, let x := x(x2

0−xTNxN )1/2

, s := s(s20−sTNsN )1/2

so that det(x) = det(s) = 1. Set γ := 1√2(1+xT s)

and w := γ(x + s −1).

It follows that also det(w) = 1, and thus, w1/2 = 1√2(1+w0)

(1 + w0

wN

). Let

η :=(

det(x)det(s)

)1/4

and

W := P ((ηw)1/2) = η

(w0 wT

N

wN I +wN wT

N

1+w0

)(20)

8

Page 9: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

= η

{(−1

I

)+

1

1 + w0

(1 + w0

wN

)(1 + w0

wN

)T}.

Since det(w) = 1 it follows from (19) that

W−1 = P ((ηw)−1/2) =1

η

(w0 −wT

N

−wN I +wN wT

N

1+w0

).

Using the definition of W one can derive the scaling condition

Ws = W−1x or equivalently, W 2s = P (ηw)s = x. (21)

The first relation in (21) can be verified for the case that det(x) = det(s) = 1– with somewhat tedious but straightforward calculations – and then, based onthe definition of η, for the general case. The second relation follows from (19).The matrix W in (21) thus satisfies the same relation as D in (7). For thisreason ηw is called the scaling point of x and s (in this order), see e.g. [19].

For general x, s ∈ SOC◦ since W is a quadratic representation of the form(16) it follows

u ◦ v = µe

⇐⇒ u = µv−1

⇐⇒ W−1u = µW−1v−1 = µ(Wv)−1

⇐⇒ (W−1u) ◦ (Wv) = µe,

where (we did not use associativity and where) the second equation in line 3uses the key relation (18). Thus, for given x, s ∈ SOC◦, the condition of findingcorrections ∆x,∆s satisfying the complementarity relation

(x+ ∆x) ◦ (s+ ∆s) = µe

can be expressed in the form

(W−1(x+ ∆x)) ◦ (W (s+ ∆s)) = µe

or

(W−1x)◦(Ws)+(W−1∆x)◦(Ws)+(W−1x)◦(W∆s)+(W−1∆x)◦(W∆s) = µe.

Let v := W−1x = Ws. Replacing v into the above equation leads to

(W−1∆x) ◦ v + v ◦ (W∆s) = µe− v ◦ v − (W−1∆x) ◦ (W∆s).

Since the (left or right) Jordan-multiplication with v is expressed by the linearoperator L(v), this is equivalent to

L(v)(W−1∆x) + L(v)(W∆s) = µe− v ◦ v − (W−1∆x) ◦ (W∆s). (22)

Now, the associativity of the matrix multiplication can be exploited by multiply-ing from left with L(v)−1. Using the identities v−1◦e = v−1 and v−1◦(v◦v) = v,it follows

W−1∆x+W∆s = µv−1 − v − L(v)−1((W−1∆x) ◦ (W∆s)). (23)

9

Page 10: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

The crucial point of the preceding derivations is that this equation can easilybe solved for ∆x or ∆s. Its derivation hinges on the scaling point ηw in (21)which was first discovered in the context of semidefinite programs5 defining theNT direction [28, 29].

3 An interior point algorithm

3.1 Search direction

In analogy to linear programs (LP ), one iteration of an interior-point methodfor solving (SOCP ) is now derived: At the start let some x, s ∈ SOC◦ be givenand some y ∈ Rm. A correction ∆x,∆y,∆s is searched for such that

A(x+ ∆x) = b

AT (y + ∆y) + (s+ ∆s) = c

W−1∆x+W∆s = µv−1 − v,

where the last line is obtained by linearizing (23) i.e. by ignoring the quadraticterm L(v)−1(W−1∆x) ◦ (W∆s).

Setting p := b−Ax, q := c−AT y− s, and r := µv−1− v = µ(Ws)−1−Ws,the above system is of the form

A∆x = p

AT ∆y + ∆s = q

W−1∆x+W∆s = r.

Solving the second and third equation for ∆s and ∆x and substituting both inthe first equation one obtains

AW 2AT ∆y = p+AW (Wq − r) (24)

∆s = q −AT ∆y (25)

∆x = Wr −W 2∆s. (26)

Note that W is a symmetric rank-1-perturbation of a diagonal matrix (as can beseen from the second line in (20)), and thus, for a dense matrix A, the evaluationof the m × n-matrix AW can be carried out with 3mn multiplications andadditions. The most expensive part of the solution of (24) is forming the matrix(AW )(AW )T with m2n/2 multiplications and additions. This is identical tothe computational effort for solving dense linear programs.

But unlike the case of linear programs, multiplication with W typically de-stroys any sparsity that may be present in A. For sparse problems, observing

5 For semidefinite programs, the scaling point of two matrices X,S � 0 is given by ηW :=W := S−1/2(S1/2XS1/2)1/2S−1/2. Set P := (W )−1/2 and replace v in (22) with V :=P−1SP−1 = PXP . Then, for some symmetric matrix A, define LV (A) := 1

2(V A + AV ).

Now, the equivalent of (22) for semidefinite programs is given by the equation

LV (P∆XP ) + LV (P−1∆SP−1) = µI − V 2 −1

2(P∆X∆SP−1 + P−1∆S∆XP ).

10

Page 11: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

that W 2 is a rank-2-perturbation of the (1 + n)× (1 + n)-matrix η2I, a sparsefactorization of AAT and a rank-2 update formula for W 2 may be used – or moregenerally, as for example in [10], a sparse LDLT factorization of an indefiniteKKT system may be used. For the more general (MSOCP ) addressed below,such rank-2 update formula is to be applied to a sparse factorization of AD2AT

for some diagonal matrix D; and when the number pS of second order cones issmall such update preserves sparsity in the same fashion as for linear programs.

We may now return to relation (10). Based on (10) the preceding transfor-mations can also be applied, the main difference, however, being that then, thematrix W 2 in AW 2AT in (24) is to be replaced with a nonsymmetric matrixmaking the numerical computations about twice as expensive. For semidefiniteprograms, the resulting search direction is called AHO direction, [2]. Apart fromthe higher computational effort, under standard nondegeneracy assumptions theAHO direction is defined in a neighborhood of the optimal solution (while theNT direction is defined only for positive definite iterates) and in numericalimplementations it returned fairly good results in terms of overall number of it-erations see e.g. [37], but its theoretical properties in the interior-point contextare slightly weaker than those of the NT direction, see e.g. [26, 37].

3.2 Step length

Equations (24) – (26) define some search direction. Given x ∈ SOC◦ and asearch direction ∆x, it is straightforward to compute the maximum step lengthαmaxx > 0 such that x+ α∆x ∈ SOC◦ for all α ∈ [0, αmax

x ):

1. If ∆x0 ≥ ‖∆xN‖, then, because SOC is a convex cone, αmaxx =∞.

2. Else, let the determinants δ∆x := ∆x20−∆xTN∆xN and δx := x2

0−xTNxN >0, and the mixed term δm := x0∆x0 − xTN∆xN be given. For δ∆x 6= 0 itfollows from 0 = det(x+ α∆x) = δx + 2αδm + α2δ∆x at the boundary ofSOC that

αmaxx =

−1

δ∆x

(δm ±

√δ2m − δ∆xδx

)(27)

where the sign of “±” is to be chosen such that αmaxx > 0, and when both

are positive, then such that αmax is the smaller of both numbers.

3. For very small |δ∆x| there may be a cancellation error in evaluating (27).

In this case consider first the sub-case that δm = 0:

If δ∆x ≥ 0, then αmaxx =∞, and if δ∆x < 0, then,

αmaxx =

√δx/|δ∆x|.

Now assume that δm 6= 0:

When δm > 0 and δ∆x < 0, the “+”-sign in formula (27) is to be taken,and then, there is no cancellation error so that the numerical computationis sufficiently stable.

If δm > 0 and δ∆x ≥ 0 then αmaxx =∞.

11

Page 12: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

When δm < 0, the Taylor approximation of the square root in (27) is usedfor numerical stability, leading to

αmaxx ≈ δx

2|δm|,

where this formula is exact for δ∆x = 0.

3.3 A general mixed linear second order cone problem

To state the general algorithm, consider now a general (MSOCP)

(MSOCP ) min { cTx | Ax = b, x ∈ K },

where K is the Cartesian product of a nonnegative orthant and of several secondorder cones. (We do not consider free variables here.) More precisely, assumethat there are pN nonnegative variables and pS second order cones of differentdimensions, i.e. K = RpN

+ × SOC(1) × . . . × SOC(pS). Let pK := pN + pS and

n = pN + dim(SOC(1)) + . . .+ dim(SOC(pS)), where dim(SOC(i)) denotes thedimension of SOC(i).

The matrix W from above is then replaced with the block diagonal matrixhaving a pN × pN diagonal block of the form D from (7) for the nonnegativevariables and diagonal blocks of the form (20) for each SOC(i).

Mehrotra’s predictor-corrector algorithm [23] applied to (MSOCP ) can thenbe stated as follows:

MPC-MSOCP-Algorithm:

(Mehrotra’s Predictor-Corrector Algorithm for solving (MSOCP ))

Input: (x0, y0, s0) with x0, s0 ∈ K◦, ε > 0, and M � 0. Set k := 0.

1. Set (x, y, s) := (xk, yk, sk), µk := (xk)T sk/pK.

2. If ‖Ax− b‖ < ε, ‖AT y + s− c‖ < ε, and µk < ε/pK: STOP.

3. If ‖x‖ > M or ‖s‖ > M : STOP.

4. Solve system (24) – (26) with µ = 0 in the definition of r. Denote thesolution by ∆xN , ∆yN , and ∆sN .

5. Find the maximum feasible step lengths along ∆xN and ∆sN ,

αNmax,x := min{1, min

i∈{1,...,pN}:∆xNi <0− xi

∆xNi, mini=1,...,pK

αmaxx,i },

αNmax,s := min{1, min

i∈{1,...,pN}:∆sNi <0− si

∆sNi, mini=1,...,pK

αmaxs,i },

where the maximum step lengths αmaxx,i , αmax

s,i with respect to SOC(i) arecomputed as in (27).

6. Setµ+N = (x+ αN

max,s∆xN )T (s+ αN

max,s∆sN )/pK

andµC = µk · (µ+

N/µk)3.

12

Page 13: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

7. Solve system (24) – (26) again (with the factorization of Step 4.) and withr := µCv

−1 − v − L(Ws)−1((W−1∆xN ) ◦ (W∆sN )).

Denote the solution by ∆xC , ∆yC , and ∆sC .

8. Choose a damping parameter ηk ∈ [0.5, 1.0) and set

αCmax,x := min{2, min

i∈{1,...,pN}:∆xCi <0− xi

∆xCi, mini=1,...,pK

αmaxx,i },

αCmax,s := min{2, min

i∈{1,...,pN}:∆sCi <0− si

∆sCi, mini=1,...,pK

αmaxs,i },

and then set

αCx := min{1, ηkαC

max,x}, αCs := min{1, ηkαC

max,s}.

9. Set

xk+1 := xk + αCx ∆xC , (yk+1, sk+1) := (yk, sk) + αC

s (∆yC ,∆sC),

and k := k + 1, and return to Step 1.

The input ε > 0 is a stopping parameter, and M � 0 an upper bound on thenorm of the iterates. A stop in Step 2. returns an approximation to the optimalsolution of (MSOCP), and a stop in Step 3. indicates that problem (MSOCP )either has no solution or it is poorly conditioned. The evaluation of r in Step7. can be carried out with O(n) arithmetic operations using the formula (12).

A complexity analysis of several related methods with a feasible startingpoint is given, for example in [36, 24]. For a method with long steps and withsearch directions as used in the above algorithm, the computational effort ofreducing the parameter µ by a factor of ε ∈ (0, 1) is O(pK log(1/ε)) Newtonsteps. Assuming exact arithmetic, this estimate is independent of any conditionnumbers; an error message that the problem may be poorly conditioned (asin Step 3. above) will not occur. When the starting point is not feasible, thecomplexity estimate may worsen, and for the predictor-corrector method above,the choice of the step lengths may be too aggressive to allow for a polynomialcomplexity estimate. In particular, a stop in Step 3. of the method may bepossible.

4 Considering an SQP subproblem

Let twice differentiable functions f : Rn → R, fE : Rn → RmE , and fI : Rn →Rmi be given and assume a problem of the form

min { f(x) | fE(x) = 0, fI(x) ≤ 0 } (28)

is to be solved. Assume further that a current iterate x is given along withsome approximate Hessian H of the Lagrangian. Let g be the gradient of fat x, and DfE be the Jacobian of the equality constraints in (28) evaluatedat x and fE be the function value of the equality constraints evaluated at x.(It shall always be evident from the context whether fE refers to the vector offunction values at x or to the function: Rn → RmE ). Let fI and DfI be given

13

Page 14: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

likewise. The Celis-Dennis-Tapia Euclidean norm trust region SQP subproblem[7] termed “CDT subproblem” then takes the form:

min gT ∆x+ 12∆xTH∆x

∆x ∈ Rn s.t.

DfE∆x − sE = −fE ,DfI∆x − sI ≤ −fI ,∥∥∥∥(sEsI

)∥∥∥∥ ≤ δC ,

‖∆x‖ ≤ δ.

(29)

where δC is the optimal value of

min t

∆x ∈ Rn s.t.

DfE∆x − sE = −fE ,DfI∆x − sI ≤ −fI ,∥∥∥∥(sEsI

)∥∥∥∥ ≤ t,

‖∆x‖ ≤ 0.9 δ.

(30)

The constant “0.9” in the last line of (30) can be replaced by some parameterρ ∈ [ρ1, ρ2] where 0 < ρ1 ≤ ρ2 < 1 are some fixed values, see e.g. [33]. Forsimplicity we state a fixed parameter “0.9” in (30).

Problems of the form (29), (30) have been proposed and analyzed for equalityconstrained problems in [7, 33]. The numerical computation of the solution ofthe SQP subproblem (for equality constrained problems) is discussed in [42, 45].The papers [8, 31, 41] analyze optimality conditions for the nonconvex equalityconstrained case. In [1, 35] semidefinite relaxations (including the inequalityconstrained case) are considered. A recent survey on this topic is given in [44].

The present paper does not consider (computationally expensive) semidefi-nite relaxations as in [1, 35] but does consider the inequality constrained caseas well for which MSOCPs offer a rather convenient tool if convexity can beenforced.

We note that the nonconvex case including inequality constraints is consid-erably more difficult than the equality constrained case. In fact the nonconvexSQP subproblem (29) in the inequality constrained case is NP-hard:

Lemma 1Problem (29) is NP-hard.

Proof:The proof is straightforward. First recall that the Max-Cut-Problem is NP-complete, see e.g. [16]. The Max-Cut problem can be written as a {−1, 1}n-quadratic program see e.g. [32]. While the optimal value and the multiplierschange, the (local or global) optimal solutions remain invariant when a multipleof the identity is subtracted from the Hessian of the quadratic objective function.After subtracting a suitable multiple (determined by Gershgorin’s theorem), theobjective function is strictly concave, so that any local solution of the (contin-uous) [−1, 1]n- quadratic program lies at a vertex and thus coincides with thesolution of the {−1, 1}n- quadratic program. The SQP subproblem (29) (withthe exact Hessian of the Lagrangian about the point x = 0, with a sufficiently

14

Page 15: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

large trust region radius δ ≥√n, and with δC = 0) for such continuous [−1, 1]n-

quadratic program is the [−1, 1]n- quadratic program itself, proving the claimof the lemma. #

Lemma 1 underlines a contrast to equality constrained problems for whichpolynomial solvability of the CDT subproblem is shown in [4] also in the non-convex case. Lemma 1 – and the fact that nonconvex SQP subproblems maygenerate search directions that neither improve feasibility nor optimality (seee.g. [17]) – provides a practical justification for the following further assumption,

H � 0 (i.e. H is positive definite),

which allows for a polynomial time solution of the SQP subproblems (29)6.In particular, for problems where the evaluation of the function values and

derivative information is computationally expensive, second order cone pro-grams offer an elegant way for solving the subproblems (29) at a predictablecomputational cost.

Second order cone programs also allow the computation of a second ordercorrection step and a re-evaluation of the Lagrange multipliers – the multipliersobtained from (29) do not approximate the Lagrange multipliers of (28) whenthe trust region constraint ∆x ≤ δ is active.

In the following we address two alternatives of converting problem (29) toa MSOCP. The first approach introduces two additional scalar variables andimplicitly makes use of a “rotated second order cone” (as introduced e.g. in[34]) while the second approach only introduces one additional scalar variableand may seem more evident and natural at a first glance. In exact arithmeticboth alternatives generate the same search directions.

First approach

As observed for example, in [6], Section 9.4, minimizing t − 1 + gT ∆x subjectto the constraints t

t− 2√2H1/2∆x

∈ SOCis equivalent to minimizing gT ∆x + 1

2∆xTH∆x. Indeed, the above SOC con-straint implies

t2 ≥ (t− 2)2 + 2‖H1/2∆x‖2, t ≥ 0,

which is equivalent to4t− 4 ≥ 2∆xTH∆x

6Enforcing this assumption may result in some loss in the quality of the solution of theSQP subproblem. In the case of a dense matrix H it can be implemented in a straightforwardfashion by first identifying inequality constraints which are anticipated as strongly active,i.e. which have a positive estimate of the Lagrange multiplier at x. If H is not positivesemi-definite, then replace H with its projection onto the null space of the gradients of theequality constraints and of the strongly active inequality constraints. Thereafter, regularizeby replacing H with H := H + θI (for some not too large θ) such that the result is positivedefinite.

If the active constraints are estimated correctly, the projection of H onto the null spacedoes not have any effect on the location of the optimal solution. If the trust region constraintis active at the optimal solution of the regularized problem (which is more likely, when θ issmall), then adding θI does not change the optimal solution either.

15

Page 16: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

or to

t− 1 ≥ 1

2∆xTH∆x.

For completeness, the reformulation of (29) to a MSOCP based on this approachis detailed next. For this reformulation denote the symmetric square root ofH � 0 by H1/2, identify t ≡ t2 and t − 2 ≡ t3. Adding nonnegative slackvariables sI to the second constraint in (29), and adding two further scalarfixed variables t0, t1 converting the Euclidean norm terms of (29) into SOC-constraints then leads to the equivalent problem

minimize gT ∆x+ 12 (t2 + t3)

s.t.

DfE∆x − sE = −fE ,DfI∆x − sI + sI = −fI ,

t0 = δ,t1 = δC ,

t2 − t3 = 2,√2H1/2∆x − x(1) = 0,

sI ≥ 0,

(t0

∆x

)∈ SOC,

t1sEsI

∈ SOC, t2t3x(1)

∈ SOC.

(31)

where “equivalent” is understood in the sense that the ∆x-part of the optimalsolution also solves (29), but the objective values differ, in general.

Note that all variables of (31) are either nonnegative or in a SOC, and thus,(31) has exactly the format (MSOCP ). Moreover, as only three second ordercones are involved, sparsity can be exploited in the same fashion as for interior-point methods for linear programs if H1/2, DfE , DfI , and A are sparse. (Ofcourse H1/2 can also be replaced with some other factor R, RTR = H thatpossibly better preserves sparsity.)

While it is possible to retrieve not only the optimal solution of (29) from(31) but also the Lagrange multipliers (using suitable rescaling), the Lagrangemultipliers only have limited relevance for problem (28); they also depend on the

first two SOC constraints of (31) (namely

(t0

∆x

)∈ SOC,

t1sEsI

∈ SOC) that

have no correspondence in (28). More appropriate estimates for the Lagrangemultipiers of (28) can be generated from the solution of an associated MSOCP.

Second approach

As an alternative to the first approach above (see e.g. [22], equation (5)), theobjective function term gT ∆x+ 1

2∆xTH∆x of (29) can be written in the form

(H−1/2g)TH1/2∆x+1

2(H1/2∆x)T (H1/2∆x) =

1

2‖g +H1/2∆x‖2 − 1

2‖g‖2

where g = H−1/2g. Since the constant term “ 12‖g‖

2” does not influence the opti-

mal solution ∆x, the objective function term can be replaced with ‖g +H1/2∆x‖2

16

Page 17: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

or with an artificial variable t ∈ R and the additional constraints(tx(1)

)∈ SOC, (32)

H1/2∆x− x(1) = −g. (33)

Again, the objective value changes with this transformation, but the optimalsolution ∆x does not.

For the second approach, the variables t2, t3 of (31) are replaced with avariable t, the fifth equation “t2 − t3 = 2” of (31) is deleted and the sixthequation of (31) is replaced with (33). The objective function is the variable t,and the last second order cone constraint in (31) is replaced with (32).

The second approach having one less variable may appear more naturalat first glance, but as shown below, it is considerably more sensitive to ill-conditioning and rounding errors. We therefore do not present a completestatement of the SQP subproblem for the second approach.

As noted by Yin Zhang [45] with regards to equality constrained problems,“As far as efficiency is concerned, we believe that methods of any kind . . . forsolving the n-dimensional Euclidean norm CDT problem are not likely to becheap . . . On the other hand, solving the CDT problem supposedly producesbetter iterative steps and therefore enhances robustness and global convergence.. . . ” As pointed out in Lemma 1 above, the problem involving inequalityconstraints is somewhat more involved than the situation referred to in [45],so that solving the convexified inequality constrained problem via an interiorpoint method for MSCOPs at a moderate computational may be suitable forproblems with expensive evaluations of g,DfE , or DfI .

5 Implementational aspects

For small scale dense problems (number of variables plus number of constraintsup to 1000), a stand-alone implementation of the MPC-MSOCP-algorithm ofthe present paper is made available for general use at the site [20]. Given thedata to problem (29) an (MSOCP ) of the form (31) is formed and its primal-dual solution is approximated with an interior-point method as in Section 3.3. Inthe end, the dual variables of (31) are rescaled and an approximate primal-dualsolution of problem (29) is generated.

5.1 Standard enhancements

Interior-point methods are known to be sensitive to ill-conditioning and to scal-ing of the data. To reduce this sensitivity, the algorithm includes some standardtechnical details not addressed in the conceptual presentation in Section 3.3, andsome of which are suitable only for dense problems:

Prior to solving problem (MSOCP ), linearly dependent rows are eliminated,a Cholesky factor L with LLT = AAT is computed and [A, b] is replaced withL−1[A, b]. (After this transformation A has orthonormal rows.) Then, c isreplaced with c − ATAc, and b and (new) c are scaled to Euclidean norm 1effecting that the origin has Euclidean distance 1 from the set {x | Ax = b},and from the set {(y, s) | AT y + s = c}. (The case b = 0 cannot occur for the

17

Page 18: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

trust region subproblem since the trust region radius δ is positive, and whenc = 0 the current iterate either is a KKT point, or one may replace c with e togenerate a direction that reduces infeasibility.) Starting points are x = s = eand y = 0. A phase 1 of several centering steps aims at reducing the violation ofthe primal-dual equality constraints while keeping µ ≡ 1. One step of iterativerefinement for the solution of the linear systems is included when the back solvereturns low accuracy results, and a safeguard is added in case that the Mehrotrapredictor-corrector strategy does not result in a sufficient decrease of µ or in casethat µ� max{‖Ax− b‖, ‖AT y + s− c‖}.

A“geometric stopping test” is used requiring the Euclidean distance to theprimal-dual equality constraints to be at most ε, i.e., ‖Ax− b‖ ≤ ε and ‖AT y+s− c‖ ≤ ε, relying on the above rescaling. The geometric stopping test furtherrequires the cosine of x and s to be less than ε,

xT s ≤ ε‖x‖‖s‖.

This stopping test is motivated by the Euclidean scaling of the starting pointand the desire of bounding the condition number of the linear systems to besolved at each iteration, see e.g. [21]. It implies, however, that in case of poorlyscaled problems (and for problems that do not have a finite optimal solution)the cosine may be small before the duality gap is below a desired threshold.

In spite of the standard safeguards as outlined above, for sparse problems –and also for poorly conditioned problems – much more sophisticated and efficientimplementations of SOCP-solvers such as SDPT3 [38] or ECOS [10] should beused; the implementation in [20] is intended as a simple stand-alone alternativefor dense SQP subproblems.

5.2 Reducing rounding errors

Interior-point algorithms typically stop when the residuals and the complemen-tarity gap are below a given threshold. Here, the norm of the residuals iscompared to the norm of the right hand side. (In Section 5.1, for example, theright hand side is first scaled to norm 1.) If certain components of the righthand side are large compared to others, then the relative accuracy with respectto the small components may be low. For a composite right hand side as in(31) this might imply that the solution found is not very accurate. We list threecases, where this effect may reduce the accuracy of an approximate solution of(31) obtained from an interior point solver and discuss how to reduce this effect.

1. Recall that g = H−1/2g enters the right hand side of the equality con-straints of the reformulation of (31) via (33) in the “second approach”.The fact that the second approach (32), (33) is feasible for H � 0 but notfor H � 0 hints at possible cancellation errors when H has small positiveeigenvalues. Such cancellation errors may indeed render an approximatesolution of the transformed MSOCP associated with (29) useless: Notethat the norm of g typically does not tend to zero in the final iterations ofan SQP algorithm. When H has tiny eigenvalues the norm of g = H−1/2gtypically is large. An algorithm that aims at finding an approximate so-lution with a small relative error in the right hand side may generate asolution with a larger error in the remaining components of the right hand

18

Page 19: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

side of (31). In the next section a simple example is given with a conditionnumber of 105 for the KKT conditions, with a strictly feasible iterate at adistance of 1 to the optimal solution, and a matrix H � 0 with conditionnumber 106, and where the solution of (29) via a MSOCP with accuracy10−10 (solved with both, SeDuMi [34], or with the solver in [20]) returnsan ascent direction for the objective function (due to rounding errors). Itis not known how to rescale the resulting MSOCP to eliminate such severerounding errors. The first approach uses one more scalar variable but doesnot suffer from any systematic ill-conditioning and is also applicable forthe slightly weaker assumption H � 0 (rather than H � 0).

2. Optimization problems sometimes come with large bounds such as “xi ≤1020” for some i ∈ {1, . . . , n}. Consider an SQP subproblem of the form(31) about a point x with moderate norm ‖x‖. When such bound istreated as an equality constraint with a slack variable si as in the secondrow of equations in (31) then such bound yields an entry of size about“1020” in the right hand side −fI of the equality constraints in (31). Asjust discussed above, very large components in the right hand side of (31)typically reduce the overall accuracy of the approximate solution obtainedfrom an interior-point solver. However, such loss of accuracy can easilybe avoided by eliminating all inequalities “i” of the SQP-subproblem forwhich −fi(x) > δ‖∇fi(x)‖ before solving the SQP subproblem. (By theCauchy Schwarz inequality such constraints will never be active.)

3. Similarly, as detailed with an example in the next section, when the trustregion radius δ in the right hand side of the equality constraints of (31)is very large but the remaining parts of the right hand side are small, theoverall accuracy of an approximate solution of (31) may degrade as well.Such situation might be possible in a trust region algorithm if the trustregion constraint was not active for several successive successful steps. Insuch situation it would be practical to reduce δ to a value at most 10 timesthe norm of the last SQP step. Such reduction would balance the righthand side of (31) but is unlikely to have much effect on the overall numberof iterations or on the local or global convergence analysis.

5.3 Numerical examples

To illustrate the sensitivity with respect to scaling and to give some intuitionabout the computational effort for solving CDT SQP subproblems, some nu-merical results are listed in this section.

The MCP-MSOCP-algorithm of this paper is similar to the approach usedin SPDT3 [38] while SEDUMI [34] uses a different self-dual approach. The com-parison of the MCP-MSOCP-algorithm with SEDUMI below suggests that thescaling issues are not due to the particular approach but rather to the condition-ing of the problem. This impression is also supported by the comparison belowreplacing the Euclidean norm trust region problem with an∞-norm trust regionproblem. For the latter, the variable ∆x is replaced with ∆x := ∆x − δe ≥ 0and an additional slack variable x(2) ≥ 0 with ∆x + x(2) = 2δe is introduced,and the objective function is transformed to an SOC-constraint and a linearobjective function as in (31).

19

Page 20: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

In Table 1 below, the two approaches considered in Section 4 for replacingthe quadratic objective function with a linear objective function and a secondorder cone constraint are compared. It turns out that the difference of the twoapproaches can be illustrated with a very simple instance of a quadratic programwith two unknowns. More precisely, let ρ > 0 be a fixed number and considerthe problem

minρ

2(x2

1 + x22) + 1000x1 + 0.1x2 where x ≥ 0. (34)

As initial point let x(0) := (10−4, 1)T .Since the starting point is feasible, it follows that δC = 0 in (30). Let δ be

given as some value δ ≥ 1.1 Then, for 0 ≤ ρ ≤ 0.1, the SQP correction in exactarithmetic is

∆x∗ = −x(0).

In Table 1 below we report the results of solving the SQP subproblem with thefirst approach, with the second approach (both using the default settings of theMehrotra predictor corrector method described in Section 3.3), and with thesecond approach using SeDuMi [34] with its default settings. The three valuesρ = 10−2, ρ = 10−4, and ρ = 10−6 are compared for the above problem. Notethat H−1/2 used in the right hand side of (33) is given by H−1/2 = ρ−1/2I. (Alsorecall that for all three values of ρ the optimal solution of the SQP subproblemremains the same.)

For each case, the norm ‖∆x − ∆x∗‖ is listed, where ∆x refers to the ap-proximate solutions generated by the three algorithms.

approach ρ = 10−2 ρ = 10−4 ρ = 10−6

first approach 6.7e-8 6.1e-9 1.3e-8second approach 2.3e-3 9.6e-3 5.2

second appr., SeDuMi 4.5 e-3 4.3 4.3

Table 1, norm of the error ∆x−∆x∗ for (34).

The numbers in Table 1 make clear that the second approach is not reliable,independent of whether the MSOCP is solved via SeDuMi or with an imple-mentation as in Section 3.3. A closer look at the results of the second approachfor ρ = 10−6 shows that the MSOCP to be solved has a right hand side of norm1.0e6. The primal residual of the solution generated by SeDuMi has a normof 1.7e-4 (almost 10 digits relative accuracy), the primal solution returned bySeDuMi has norm 1.4e6, the ∆x-part of this solution has norm 3.3e0. Evenfor this simple problem, looking at the solution generated by SeDuMi, it is notclear, how to rescale the problem to reduce rounding errors; except from usingthe first approach.

Table 2 refers to 100 randomly generated problems with 200 variables each,100 equality constraints and 200 inequality constraints. The objective term gand 100 entries of fI were generated uniformly distributed in [0, 1], 50 entries offI were generated uniformly distributed in [−1, 0] and the remaining entries offI were set to zero. All the remaining data for Table 2 was generated standardnormally distributed. The matrix H generated this way was multiplied by itstranspose to make it positive definite. To test the effect of an overly large trustregion radius δ, an initial value δ := 104 was chosen which happened to be

20

Page 21: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

inactive for all 100 test problems. Then, a second run was performed for whichδ was reduced to twice the norm of the solution ∆x generated in the first run (tomake sure the trust region constraint is still inactive). The average reductionof δ was by a factor 1.4e-3. Table 2 lists the smallest and the largest norm ofthe violation of the KKT conditions before and after reducing δ, as well as thenorm of the constraint violation before and after reducing δ.

result \ approach MCP-MSOCP SEDUMI

‖KKTviol.‖, before 1.2e-3 2.0e-3

‖KKTviol.‖, after 2.1e-5 1.2e-4

‖feas.viol.‖, before 1.0e-2 1.4e-4

‖feas.viol.‖, after 4.4e-7 7.3e-5

Table 2, average KKT / constraint violation before and after reducing δ.

The examples generated for Table 2 resulted in a solution ∆x for which thenorm of H∆x typically was much larger than the norm of g, accounting for acancellation error in the evaluation of the KKT conditions. The crucial part forthese problems, namely the violation of feasibility, is reduced effectively when δis small.

To obtain some intuition about the running times for solving Euclideannorm subproblems compared to ∞-norm subproblems, feasible random sparsesubproblems were generated and tested with SEDUMI. For feasible problemsδC = 0 and the second SOC constraint in (31) (involving the variable t1) iseliminated prior to solving the MSOCP. The problems included n variables,n/2 equality constraints and n inequality constraints. The average runningtimes over 10 random problems are listed in Table 4:

approach \ n 20 200 2000Euclidean Norm 0.13 0.24 171∞-Norm 0.10 0.29 226

Table 3, solution times in seconds

For the l2 problem the dimension n = 20000 was tested as well. This dimen-sion resulted in an MSOCP with 60000 variables and 50000 equations. (For thel∞ problem there are 70000 variables and 60000 equations.) The matrix A of theMSOCP had 0.022 percent nonzero entries, the factor ADAT had 0.79 percentnonzero entries and its Cholesky factor had 18 percent nonzero entries (i.e. over400 Million nonzeros) after eliminating 3 dense columns from A. The numbersindicate that the approach is not readily transferrable to large scale problems.While large scale problem problems would need to be solved by different ap-proaches the results do seem to indicate that there is no loss in efficiency whensolving Euclidean norm trust region subproblems with interior point approachesrather than infinity norm trust region subproblems. The same would be truefor the comparison of Euclidean norm trust region subproblems with one-normtrust region subproblems (having one additional equation compared to infinitynorm trust region subproblems).

21

Page 22: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

6 Summary

While the theory of interior-point methods for linear and semidefinite programsis well established and implementations are used for many practical applications,there still seems to be a lack regarding the general acceptance of second ordercone programs. To lessen the reservations toward SOCPs, in chapter 2 and 3, aself-contained presentation of practical aspects of an interior point method forsolving SOCPs is given. One important application where SOCP can be usefulare trust region SQP subproblems. This application is addressed in detail andsome scaling/conditioning parameters that influence the numerical accuracy ofan approximate solution are analyzed. Preliminary numerical examples illus-trate the effects of the scaling parameters. As a future work the approach is tobe used in an SQP method using finite differences for the approximations of thederivatives.

Acknowledgment

The author likes to thank two anonymous referees and the editors who helpedto extend and to improve this paper.

References

[1] Ai, W.B., Zhang, S.Z.: “Strong duality for the CDT subproblem: a nec-essary and sufficient condition”, SIAM J. Optim. 19, 1735–1756 (2009).

[2] Alizadeh, F., Haeberly, J.-P., and Overton, M.: “Primaldual interior-pointmethods for semidefinite programming: convergence rates, stability andnumerical results”, SIAM J. Opt. 8, 746–768 (1998).

[3] F. Alizadeh, F., and Goldfarb, D.: “Second-order cone programming”,Math. Prog. 95, 3–51 (2003).

[4] Bienstock, D.: “A Note on Polynomial Solvability of the CDT Problem”,SIAM J. Optim., 26(1), 488–498 (2016)

[5] Boyd, S.; Vandenberghe, L.: Convex Optimization (PDF), Cambridge Uni-versity Press (2004).

[6] Calafiore, G.C., El Ghaoui, L.: “Optimization Models”, Cambridge Uni-versity Press, Control systems and optimization series, (2014)

[7] Celis, M., Dennis, J., Tapia, R.: “A trust region strategy for nonlinearequality constrained optimization”, Numerical Optimization (P. Boggs, R.Byrd, and R. Schnabel, eds), SIAM, 71–82 (1985).

[8] Chen, X.D., Yuan, Y.: “On local solutions of the CDT subproblem”, SIAMJ. Optim. 10, 359–383 (1999).

[9] Conn, A. R. , Gould, N. I. M., Toint, Ph. L.: “Trust-region methods”,Society for Industrial and Applied Mathematics, Philadelphia, PA, USA,(2000).

22

Page 23: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

[10] Domahidi, A., Chu, E., and Boyd, S.: “ECOS: An SOCP Solver for Embed-ded Systems”, Proceedings European Control Conference, pp 3071–3076,Zurich, Switzerland, July 2013.

[11] Faraut, J., and Koranyi, A.: Analysis on symmetric cones. Oxford Mathe-matical Monographs, The Clarendon Press Oxford University Press, OxfordScience Publications, New York, (1994).

[12] Faybusovich, L.: “Linear systems in Jordan algebras and primal-dualinterior-point algorithms”, Journal of Computational and Applied Math-ematics 86 (1), 149–175 (1997).

[13] Faybusovich, L.: “Euclidean Jordan Algebras and Interior-point Algo-rithms”, Positivity Volume 1, Issue 4, 331–357 (1997).

[14] Fletcher, R.: “A model algorithm for composite non-differentiable opti-mization problems”, Math. Programming Studies 17, 67–76 (1982).

[15] Fletcher, R., Leyffer, S.: “Nonlinear programming without a penalty func-tion”, Math. Prog. 91, 239–269 (2002).

[16] Garey, M. R., Johnson, D. S.: “Computers and Intractability: A Guide tothe Theory of NP-Completeness”, A Series of Books in the MathematicalSciences. San Francisco, USA, W. H. Freeman and Co. (1979).

[17] Goldsmith, M.: “Sequential Quadratic Programming Methods Based on In-definite Hessian Approximations”, PhD thesis, Stanford University (1999).

[18] Gould, N. I. M.: “An introduction to algorithms for continuous optimiza-tion”, Report, Oxford University Computing Laboratory and RutherfordAppleton Laboratory,http://www.cs.ox.ac.uk/nick.gould/courses/co/paper.pdf (2006).

[19] Gu, G., Zangiabadi, M., Roos, C.: “Full Nesterov-Todd step interior-pointmethods for symmetric optimization”, Eur. J. Oper. Res. 214(3), 473–484(2011).

[20] Jarre, F.: MWD, smooth Minimization Without using Derivatives, a Mat-lab collection.http://www.opt.uni-duesseldorf.de/en/forschung-fs.html (2015).

[21] Jarre, F.: “A priori bounds on the condition numbers in interior-pointmethods”,http://www.optimization-online.org/DB HTML/2015/08/5052.html (2015).

[22] Lobo, M.S., Vandenberghe, L., Boyd, S., Lebret, H.: “Applications of sec-ond order cone programming”, Lin. Alg. Appl. 284, 193–228 (1998).

[23] Mehrotra, S.: “On the implementation of a primal-dual interior pointmethod”, SIAM Journal on Optimization 2 (4), 575–601 (1992).

[24] Monteiro, R. D. C., and Tsuchiya, T.: “Polynomial convergence of primal-dual algorithms for the second-order cone program based on the MZ-familyof directions”, Math. Prog. 88, 61–83 (2000).

23

Page 24: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

[25] More, J.J., Sorensen, D.C.: “Computing a trust region step”, SIAM J. Sci.Statist. Comput. 4, 553–572 (1983).

[26] Muramatsu, M., Vanderbei, R.J.: “Primal-Dual Affine-Scaling AlgorithmsFail for Semidefinite Programming”, Mathematics of Operations Research24(1), 149–175 (1999).

[27] Nesterov Y.E., and Nemirovskii, A.S.: Interior-Point Polynomial Algo-rithms in Convex Programming, SIAM Studies in Applied Mathematics,SIAM, Philadelphia (1993).

[28] Nesterov, Y.E., and Todd, M.J.: “Self-Scaled Barriers and Interior-PointMethods for Convex Programming”, Math. of OR 22, 1–42 (1997).

[29] Nesterov, Y.E., and Todd, M.J.: “Primal-dual interior-point methods forself-scaled cones”, SIAM Journal on Optimization 8(2), 324–364 (1998).

[30] Omojokun, E.O.: “Trust region algorithms for optimization with nonlinearequality and inequality constraints”, Doctoral Dissertation, University ofColorado at Boulder Boulder, CO (1989).

[31] Peng, J., Yuan, Y.: “Optimality conditions for the minimization of aquadratic with two quadratic constraints”, SIAM J. Optim. 7, 579–594(1997).

[32] Poljak, S., Rendl, F.: “Nonpolyhedral Relaxations of Graph-BisectionProblems”, SIAM J. Optim., 5(3), 467–487 (1995).

[33] Powell, M.J.D., Yuan, Y.: “A trust region algorithm for equality con-strained optimization”, Math. Program. 49, 189–211 (1991).

[34] Sturm, J.F.: “Using SeDuMi 1.02, a MATLAB toolbox for optimizationover symmetric cones”, Optimization Methods and Software 11-12, 625–653(1999).

[35] Sturm, J.F., Zhang, S.Z.: “On cones of nonnegative quadratic functions”,Math. Oper. Res. 28, 246–267 (2003).

[36] Tsuchiya, T.: “A convergence analysis of the scaling-invariant primal-dualpathfollowing algorithms for second-order cone programming”, Optimiza-tion Methods and Software 11 and 12, 141-182 (1999).

[37] Tutuncu, R.H., Toh, K.C., and Todd, M.J.: “On the Nesterov–Todd Direc-tion in Semidefinite Programming”, SIAM Journal on Optimization 8 (3),769–796, (1998).

[38] Tutuncu, R.H., Toh, K.C., and Todd, M.J.: “Solving semidefinite-quadratic-linear programs using SDPT3”, Mathematical Programming, Ser.B, 95, 189–217 (2003).

[39] Vardi, A.: “A trust region algorithm for equality constrained minimization:convergence properties and implementation”, SIAM J. Num. Anal. 22, 575–591 (1985).

[40] Wright, S.J.: Primal-Dual Interior-Point Methods SIAM, Philadelphia(1997).

24

Page 25: The Solution of Euclidean Norm Trust Region SQP ...The presentation is in- ... November 11, 2016 1. 1 Introduction The trust region subproblem for unconstrained minimization typically

[41] Yuan, Y.: “On a subproblem of trust region algorithms for constrainedoptimization”, Math. Program. 47, 53–63 (1990).

[42] Yuan, Y.: “A dual algorithm for minimizing a quadratic function with twoquadratic constraints”, J. Comput. Math. 9, 348–359 (1991).

[43] Yuan, Y.-X.: “On the convergence of a new trust region algorithm”, Numer.Math. 70, 515–539 (1995).

[44] Yuan, Y.: “Recent advances in trust region algorithms”, Math. Program.151, 249–281 (2015).

[45] Zhang,Y.: “Computing a CelisDennisTapia trust region step for equalityconstrained optimization”. Math. Program. 55, 109–124 (1992).

25