a novel three-phase trajectory informed search methodology for

17
J G lob O ptim (2007) 38:6177 D O I 1 0 . 1 0 0 7 / s 1 0 8 9 8 - 0 0 6 - 9 0 8 3 - 3 O R I G I N A L A R T I C L E A n o v e l t h r e e - p h a s e t r a j e c t o r y i n f o r m e d s e a r c h m e t h o d o l o g y f o r g l o b a l o p t i m i z a t i o n J a e w o o k L e e R e c e i v e d : 2 9 J u n e 2 0 0 6 / A c c e p t e d : 7 A u g u s t 2 0 0 6 / P u b l i s h e d o n l i n e : 6 S e p t e m b e r 2 0 0 6 ' S p r i n g e r S c i e n c e + B u s i n e s s Me d i a B . V . 2 0 0 6 Abstract A n ew deterministic method for s olving a g lobal optimization p roblem is p r o p o s e d . T h e p r o p o s e d m e t h o d c o n s i s t s o f t h r e e p h a s e s . T h e r s t p h a s e i s a t y p i - c a l l o c a l s e a r c h t o c o m p u t e a l o c a l mi n i m u m . T h e s e c o n d p h a s e e m p l o y s a d i s c r e t e sup-local s earch to locate a s o-called s up-local minimum taking the l owest objective v a l u e a m o n g t h e n e i g h b o r i n g l o c a l m i n i m a . T h e t h i r d p h a s e i s a n a t t r a c t o r - b a s e d g l o b a l s e a r c h t o l o c a t e a n e w p o i n t o f n e x t d e s c e n t w i t h a l o w e r o b j e c t i v e v a l u e . T h e simulation results t hrough w ell-known g lobal optimization p roblems a re shown t o d e m o n s t r a t e t h e e f c i e n c y o f t h e p r o p o s e d m e t h o d . K e y w o r d s Global optimization D y n a m i c a l s y s t e m s Deterministic methods 1 I n t r o d u c t i o n C o m p u t a t i o n o f t h e g l o b a l o p t i m a l s o l u t i o n s f o r n o n - l i n e a r p r o g r a m m i n g i s i m p o r t a n t i n a v e r y b r o a d r a n g e o f a p p l i c a t i o n s . A s i t s i m p o r t a n c e a l o n e m a y s u g g e s t , g l o b a l o p t i m i z a t i o n i s g e n e r a l l y a n e x t r e m e l y h a r d p r o b l e m . F r o m t h e c o m p l e x i t y p o i n t o f view , g lobal optimization p roblems b elong to the c lass of NP-hard p roblems ( [30]), w h i c h me a n s t h a t t h e c o m p u t a t i o n a l t i m e r e q u i r e d t o s o l v e t h e p r o b l e m i s e x p e c t e d t o g r o w e x p o n e n t i a l l y a s t h e i n p u t s i z e o f t h e p r o b l e m i n c r e a s e s . D e s p i t e s u c h d i f c u l t i e s , h o w e v e r , n u m e r o u s a l g o r i t h m s f o r s o l v i n g g l o b a l o p t i m i - z a t i o n p r o b l e m s h a v e b e e n d e v e l o p e d d u r i n g t h e l a s t t w o d e c a d e s d u e t o i n c r e a s i n g needs o f most e ngineers a nd scientists to solve c omplex real-world systems . Most exist- i n g me t h o d s c a n b e c l a s s i e d i n t o t w o c a t e g o r i e s a s i n [ 1 2 , 1 7 ] : s t o c h a s t i c me t h o d s a n d deterministic methods . S tochastic methods sample the objective function for a small n u m b e r o f p o i n t s a n d i n c l u d e g e n e t i c a l g o r i t h m s , s i m u l a t e d a n n e a l i n g , c l u s t e r i n g J . L e e ( B ) D e p a r t m e n t o f I n d u s t r i a l E n g i n e e r i n g , P o h a n g U n i v e r s i t y o f S c i e n c e a n d T e c h n o l o g y , P o h a n g , K y u n g b u k 7 9 0 - 7 8 4 , K o r e a e - m a i l : j a e w o o k l @ p o s t e c h . a c . k r

Upload: others

Post on 03-Feb-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77DOI 10.1007/s10898-006-9083-3

ORIGINAL ARTICLE

A novel three-phase trajectory informed searchmethodology for global optimization

Jaewook Lee

Received: 29 June 2006 / Accepted: 7 August 2006 / Published online: 6 September 2006© Springer Science+Business Media B.V. 2006

Abstract A new deterministic method for solving a global optimization problem isproposed. The proposed method consists of three phases. The �rst phase is a typi-cal local search to compute a local minimum. The second phase employs a discretesup-local search to locate a so-called sup-local minimum taking the lowest objectivevalue among the neighboring local minima. The third phase is an attractor-basedglobal search to locate a new point of next descent with a lower objective value. Thesimulation results through well-known global optimization problems are shown todemonstrate the ef�ciency of the proposed method.

Keywords Global optimization · Dynamical systems· Deterministic methods

1 Introduction

Computation of the global optimal solutions for non-linear programming is importantin a very broad range of applications. As its importance alone may suggest, globaloptimization is generally an extremely hard problem. From the complexity point ofview, global optimization problems belong to the class of NP-hard problems ([30]),which means that the computational time required to solve the problem is expectedto grow exponentially as the input size of the problem increases.

Despite such dif�culties, however, numerous algorithms for solving global optimi-zation problems have been developed during the last two decades due to increasingneeds of most engineers and scientists to solve complex real-world systems. Most exist-ing methods can be classi�ed into two categories as in [12, 17]: stochastic methods anddeterministic methods. Stochastic methods sample the objective function for a smallnumber of points and include genetic algorithms, simulated annealing, clustering

J. Lee (B)Department of Industrial Engineering, Pohang University of Scienceand Technology, Pohang, Kyungbuk 790-784, Koreae-mail: [email protected]

Page 2: A novel three-phase trajectory informed search methodology for

62 J Glob Optim (2007) 38:61–77

methods, bayesian methods, etc. This class of methods does not need derivativeinformation, but the required number of samples to arrive at a desired solution is oftennot admissible for large problems. On the other hand, deterministic methods typicallyprovide a mathematical guarantee for convergence to anε-global minimum in a �nitenumber of steps for optimization problems involving certain mathematical structures.Branch and bound methods, cutting planes, decomposition-based approaches, cover-ing methods, interval methods, interior point methods, and tunnelling methods are alldeterministic in nature.

In this paper, a new deterministic methodology for global optimization is presented.This method consists of three-phases: Phase I for approaching a new local minimumin terms of a gradient system-based local search, Phase II for locating a sup-localminimum in terms of a discrete local search, and Phase III for approaching a feasiblesolution with a lower objective value than those of previously obtained solutions interms of an attractor-based global search.

The proposed method was stimulated by the study of neural optimization network([8, 9, 29]) and recent progress in tunnelling methods ([2, 24, 31]), for example, bya code called TRUST proposed by Barhen and co-workers ([2, 7]). Their algorithm,employing a descent-tunnelling methodology characterized by non-Lipschitzian ter-minal repeller and 1-Dimensional (1-D) search, has made a signi�cant improvementin solving a variety of global optimization problems. For 1-D problems, the TRUSTalgorithm is designed to sweep the whole search space and thus �nd the global min-imum. However, a direct extension of the 1-D scheme to multi-dimensional globaloptimization problems does not guarantee that the global optimum will always befound since it cannot capture a trajectory that tunnels near the periphery of the basinof attraction of a global minimum if the surface gradients are weak.

Our proposed method overcomes the above weakness of the tunnelling strategyby employing two innovative ideas. First, it builds a so-called optionally scaled sys-tem, which enables an attractor-based trajectory search to capture a trajectory thattraverses near a feasible region with lower objective values. Second, it introduces anovel concept of neighboring local minima adjacent to a local minimum based ondynamical characterization, which enables a discrete sup-local search to speed up thesearch process.

2 Basic idea

A general global optimization problem is de�ned as follows:

minx

f (x) (1)

where f : D � � n � � is assumed to be smooth and the setD will be referred to asthe set of feasible points (or search space). A pointx� � D is called a local minimumif f (x� ) � f (x) for all x � D with � x Š x� � < δ for some δ > 0.

The proposed method consists of three-phases:

€ Phase I is a continuous local search and looks for a local minimum.€ Phase II is a discrete sup-local search and looks for a sup-local minimum with the

lowest objective value among the neighboring local minima.€ Phase III is an attractor-based global search and looks for a feasible solution with

a lower objective value than those values obtained in Phase II.

Page 3: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 63

0 2 4 6 8 10–2

–1

0

1

2

3

4

5

: local minima

(a) phase I

0 2 4 6 8 10–2

–1

0

1

2

3

4

5

: local minima

: suplocal minima

(b) phase II

0 2 4 6 8 10–1.5

–1

0.5

0

0.5

1

1.5

2(c) phase III

0 2 4 6 8 10–2

–1

0

1

2

3

4

5

: local minima

: suplocal minima

(d) final stage

•initial point

point found after phase I

point found after phase II

point found after phase III

starting point

starting point

point found after phase I

starting point

•point found after phase II

x0

x1

x1

x2

x2

x3

x4

x4

x5

Fig. 1 Illustration of the proposed method for 1-D case

To illustrate the basic idea of the proposed method, consider a 1-D example shownin Fig. 1. Starting from a given initial point x0, Phase I searches for a local minimum,sayx1, using a local search solver (see Fig.1a). Next, Phase II iteratively generates aset of neighboring local minima starting from x1, each of them is “locally” improvedwith respect to its neighboring local minima until it gets stuck at a locally optimalsolution, say x2, which we will call a sup-local minimum (see Fig.1b). At a sup-localminimum x2, Phase II is no longer helpful to �nd a new local optimal solution withlower objective value. Phase III is then invoked to escape from x2 and to move ontoward another point, say x3, with lower objective value than f (x2). From x3, PhaseI is invoked to �nd a local minimum, say x4, followed by Phase II, which locates asup-local minimum, say x5. In this way, the three phases are repeated alternativelyin the search space until a global minimum is found or a stopping condition is met.In the next section, we describe a detailed procedure to handle a multi-dimensionalglobal optimization problem where we speci�cally focus on the unconstrained case.

3 The proposed method

3.1 Phase I: local search phase

The aim of Phase I is to �nd a (�rst-order) local minimum. To this end, we build ageneralized negative gradient system described by

Page 4: A novel three-phase trajectory informed search methodology for

64 J Glob Optim (2007) 38:61–77

dxdt

= Š gradRf (x) � Š R(x)Š1 f (x), (2)

where R(x) is a positive de�nite symmetric matrix for all x � � n. Such anR is calleda Riemannian metric on � n [19]. It is interesting to note that many local search algo-rithms can be considered as a discretized implementation of process (2) dependingon the choice of R(x). For example, if R(x) = I, it is the naive steepest descentalgorithm; if R(x) = Bf (x) where Bf is a positive de�nite matrix approximating theHessian, 2f (x), then it is the Quasi–Newton method; if R(x) = [ 2f (x) + µI], it isthe Levenberg–Marquardt method, and so on [27]. Otherwise speci�ed, we assumethat f is twice differentiable to guarantee the existence of a unique solution (or tra-jectory) x(·): � � � n for each initial condition x(0). Note that it can be shown thatthe trajectory x(·) is de�ned on all t � � for any initial condition x(0) under a suitablere-parametrization [14].

A state vector x̄ satisfying the equation f (x̄) = 0 is called anequilibrium point (orcritical point) of system (2). We say that an equilibrium point x̄ of (2) is hyperbolicif the Hessian of f at x̄, denoted by Hf (x̄), has no zero eigenvalues (i.e.,Hf (x̄) ispositive de�nite). Note that all the eigenvalues of the Jacobian of a gradient systemare real since they are symmetric matrices. A hyperbolic equilibrium point is called a(asymptotically) stable equilibrium point (or an attractor) if all the eigenvalues of itscorresponding Jacobian are positive and anunstable equilibrium point (or a repellor)if all the eigenvalues of its corresponding Jacobian are negative. A hyperbolic equi-librium point x̄ is called anindex-k equilibrium point if the Jacobian of the hyperbolicequilibrium point has exactly k negative eigenvalues. Thebasin of attraction of a stableequilibrium point xs is similarly de�ned as

A(xs) :={x(0) � � n : lim

t�x(t) = xs

}.

From a topological point of view, the basin of attraction A(xs) is an open andconnected set.

One nice property of this formulation is that every local minimum of the opti-mization problem ( 1) corresponds to a (asymptotically) stable equilibrium point ofsystem (2), i.e., x̄ is a stable equilibrium point of ( 2) if and only if x̄ is an isolatedlocal minimum for ( 1) [19]. Hence, the task of �nding the local minima of ( 1) can beachieved via locating corresponding stable equilibrium points of (2). Another niceproperty is that all the generalized gradient system have the equilibrium points at thesame locations with the same type [19], i.e., ifR1(x) and R2(x) are Riemannian metricson � n, then the locations and the types of the equilibrium points of gradR1

f (x) andgradR2

f (x) are the same. This convenient property allows us to design a computation-ally ef�cient algorithm to �nd stable equilibrium points of ( 2), and thus to �nd localoptimal solutions of ( 1).

Phase I can be simply implemented by numerically integrating system (2) or by asteepest-descent method. However, since we are primarily concerned with �nding astable equilibrium point, which is a limit point of a trajectory, rather than the trajec-tory itself, we have the freedom to choose any robust local search algorithm if it canlocate a local minimum nearby an initial guess [5]. For example, a hybrid local searchalgorithm such as trust region-based methods, BFGS methods, or SQP methods usingsecond-order information [1, 11, 27] (or line search methods, simplex methods ortheir modi�cations when no derivative information is available [21, 26, 27]) can beemployed for computational ef�ciency. Notice that from the ‘locality’ of search in

Page 5: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 65

Phase I, to avoid jumping out of the current basin of attraction, it is important not totake too large steps initially for the line search employed by a local search method.

3.2 Phase II: discrete sup-local search phase

When we get stuck at a local minimum in Phase I, one important issue is how to escapefrom a local minimum and move on toward another neighboring local minimum. For a1-D case, a neighboring local minimum adjacent to a local minimum,xs, can be simplyde�ned as the local minima nearest to xs in each direction of � . In multi-dimensionalcases, we de�ne a neighboring local minimum adjacent to a local minimum as follows.

DeÞnition 1 Let xs be a local minimum of (1). We say that a local minimum ys is adja-cent to xs if there exists an index-one equilibrium point x1 such that the 1-D unstablemanifold Wu(x1)

1 of x1 converges to bothxs and ys with respect to system (2).

Note that such an index-one equilibrium point x1 is on ∂A(xs) � ∂A(ys). (See Fig.2.)From a theoretical viewpoint, this de�nition enables us to build a graph G = (V, E)

describing the connections between the local minima with the following elements:

1. The verticesV of G are local minima x1s, . . . , xp

s of (1), where p is the total numberof local minima

2. The edgeE of G can only connect two local minima adjacent to each other;xis is

connected with xjs if, and only if, xi

s is adjacent toxjs.

It can be shown that the graph G is connected under some mild conditions.The original optimization problem ( 1) can now be transformed into the following

combinatorial optimization problem

minxs� L1

f (xs), (3)

where L1 is the set of all the local minima.Utilizing the concept of a neighboring local minimum, Phase II uses a discrete

local search strategy onL1 to solve problem (3). Speci�cally, we �rst construct a user-de�ned neighborhood N where the neighborhood N (xs) � L1 of xs � L1 is de�ned asthe set of multiple neighboring local minima adjacent to (or near) xs. With respect tothe neighborhood N , we will call a point x� � L1 a sup-local minimum if f (x� ) � f (ys)

whenever ys � N (xs). (It should be noted that the sup-local optimality depends onthe neighborhood function that is used and the neighborhood should guarantee theconnectedness to be solvable. See Fig.3.) Then we perform a discrete local searchstrategy to locate a sup-local minimum.

There are two widely-used discrete local search strategies for solving problem (3):best-neighboring and hill-climbing, each involving a search inN (xk) of the currentxk. In a best-neighboring strategy, the point leading to the maximum decrease inthe objective function value is selected to update the current point. Therefore, allpoints in N (xk) need to be searched every time, leading to computation complex-ity of O(m), where m is number of points in N (xk). In hill-climbing, the �rst pointleading to a decrease in the objective function value is selected to update the currentpoint. Most often, best-neighboring strategies are more computationally expensivethan hill-climbing.

1 An unstable manifold Wu(x̄) is de�ned as the set of all points such that limt�Š x(t) = x̄.

Page 6: A novel three-phase trajectory informed search methodology for

66 J Glob Optim (2007) 38:61–77

–3 –2 –1 0 1 2 3–3

–2

–1

0

1

2

3

• 1

• 1 • 0• x

s

• 0

• 0

• 0

• ys

• 2

• 2

A(ys)

A(xs)

xs,y

s: stable equilibrium points

x1: index–1 equilibrium point

x1

Wu(x1)

Fig. 2 A phase portrait of the negative gradient system corresponding to a six-hump camelbackfunction ( CA). The index-1 equilibrium point x1 is a type-one equilibrium point connecting two localminima xs and ys

To locate a neighboring local minimum adjacent to a given local minimum, we mayapply the following direct strategy: in order to escape from a local minimum, say xs,we �rst move in reverse time from xs to an index-1 equilibrium point, say x1. Then,starting from x1, we advance time to approach an adjacent local minimum, sayys (seeFig. 2). The latter step can also be implemented using a local search algorithm as inPhase I. The former step, although it is hard to perform directly and time-consuming,can also be accomplished by adopting some of the techniques suggested in [19, 23].This direct strategy can, however, be inef�cient since we are only interested in locat-ing an adjacent local minimum and do not need to �nd an exact index-1 equilibriumpoint, which is computationally intensive.

The following indirect strategy can instead be used to speed up the search of neigh-boring local minima: Choose an initial straight ray r(λ), emanated from r(0) = xsand follow the ray r(λ), with increasing λ, (or employ a mixed cubic and quadraticinterpolation line search algorithm) until it reaches its �rst local minimum (or it hitsthe boundary of search space) off (r(·)). (The objective function f would increase,pass a local maximum, and reach a local minimum along the ray.) Starting from

Page 7: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 67

(a) (b)

Fig. 3 User-de�ned neighborhood structure

the obtained point, locate a neighboring local minimum using a local search algo-rithm as in Phase I. (If a neighboring local minimum cannot be found, try anotherray.) (see Fig.3).

3.3 Phase III: attractor-based global search phase

When Phase II arrives at a sup-local minimum inL2, it stops and no longer proceedsto �nd a new feasible or local optimal solution with lower objective value than thoseof previously obtained solutions. One important issue is now how to escape from asup-local minimum, sayx� , and to �nd a new point of next descent in the lower levelset given by

S(x� ) ={

x � � n : fx� (x)def= f (x) Š f (x� ) < 0

}. (4)

In general, the set S(x� ) can be decomposed into several disjoint connected compo-nents, i.e.,

S(x� ) =m⋃

j= 1

Cj,

where Cj are disjoint connected components and will be called afeasible componentof (4).

Phase III aims at locating a point in a feasible component of (4) and is based on thefollowing idea: if it is possible to �nd a nonlinear dynamical system such that all thefeasible components are attractors (whose nearby trajectories approach it asymptot-ically) of the corresponding dynamical system, then it is possible to to �nd a feasiblecomponent of (4) by �nding an attractor of the corresponding dynamical system.

Page 8: A novel three-phase trajectory informed search methodology for

68 J Glob Optim (2007) 38:61–77

To implement this idea, we build the following optionally scaled attractingdynamical system associated withfx� :

dxdt

= Š ρ(fx� (x)) f (x), (5)

where ρ(·) denotes a quasi-max zero function of one variable given by

ρ(s) ={

0, if s � 0,m(s), if s > 0

and m(s) is a strictly increasing differentiable function of s � 0 such that m(0) = 0(hence,m(s) > 0 if s > 0).

Remark 1

(1) System (5) has a Lyapunov function given by E(x) := 12ρ(fx� (x))2 since

dE(x)

dt= Š ρ(fx� (x))2� f (x)� 2 max{m (fx� (x)), 0} � 0

(2) In general, a quasi-max zero function ρ(s) need not be differentiable at s = 0,but can be approximated by a differentiable function. One possible candidatefor such a differentiable function is

ρ(s) =1α

ln(1 + exp(αm(s))) � max{m(s), 0},

whereα > 1 is a constant that affects the asymptotic behavior of the transforma-tion. This choice of ρ combined with the assumption that f is twice differentiableguarantees the existence of a unique solution (or trajectory) x(·): � � � n foreach initial condition x(0).

The next theorem establishes a result showing the relationship between feasible com-ponents of (4) and attractors of (5), which is a distinguished feature of system (5).

Theorem 1 Every feasible component of (4) corresponds to an attractor of (5), i.e., ifC is a feasible connected component of (4), then C is an attractor of system (5).

Proof See Appendix.��

Theorem 1 asserts that the task of �nding a new point in a feasible component of(4) can be achieved via locating its corresponding attractor of system (5). It shouldbe, however, pointed out that the attractors of system (5) may not always be feasiblecomponent of (4). It is due to the fact that there is a possibility that the numericalintegration of system (5) stops at a point x̄ where the Lyapunov function E(x̄) > 0.

Despite this fact, introducing system (5) endows the proposed method with a muchwider convergent regions than the tunnelling algorithm suggested in ([2, 7]). To illus-trate this, let us consider TRUST or its variants that employ a descent-tunnellingmethodology characterized by non-Lipschitzian terminal repeller and 1-D searchwhere the tunnelling is performed by integrating a dynamical system given by:

dxi

dt= (xi Š x�

i )13 θ(fx� (x)) (6)

Page 9: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 69

from the initial condition x�i + δi where θ(·) is the Heaviside function, which is equal to

1 for positive values of the argument and zero otherwise. This system dynamics trans-forms a local minimum x� into a terminal repeller x� violating the Lipschitz conditionand escapes this local minimum starting from any initial point nearby x� in a �nitetime ([2, 7]). This method, however, fails to �nd a new region of next descent unlessa system trajectory of (6) during tunnelling reaches a feasible component of (4), i.e.,inside an attractor of system (5). Therefore, a large number of search directions arerequired to �nd a proper trajectory that tunnels the region, resulting in very high-computational complexity for higher dimensional cases. In contrast, the proposedmethod can capture a trajectory outside this attractor, or more exactly in the basinof attraction of this attractor, thereby, leading the captured trajectory to a feasiblecomponent of (4) (see Fig.4).

3.4 Algorithm and implementation

We are now in a position to present a conceptual algorithm for the proposed methodas follows.

Algorithm 1 Proposed Method

1. (Initialization)1.1. Initialize x̄ and a toleranceε > 0;1.2. SetV = � ; // a set of local minima

2. (Phase I: Local Search Phase)2.1. Compute a local minimumxs starting from x̄; setV = V � { xs};2.2.if xs satis�es the termination conditions, then stop; xs is a desired solution;

3. (Phase II: Discrete Sup-Local Search Phase)do3.1. Compute a neighboring local minimumys � N (xs) of xs3.2.if f (ys) < f (xs), then set xs := ys and V := V � { xs};until no neighboring local minimum is found that improves the currentlocal minimumSetx� := xs; // x� is a sup-local minimum //

4. (Phase III: Attractor-based Global Search Phase)for i = 2 to l do4.1. choose a escaping pointxi from x�

4.2. Locate an attractor of (5) starting from xi (e.g., by numerically integrationof (5));Set the obtained point by x̄;

4.3.if f (x̄) Š f (x� ) < ε, then goto Phase I;end

Remark 2

1. Any local search algorithm designed to locate a limit point of a trajectory of ( 5)starting from xi can also be used in step 4.2 as discussed in Sect.3.1.

2. The issue of how to choose an appropriate pointxi in step 4.1 of Phase III is noteasy to address. This issue is raised since, as pointed out in the previous section, anattractor of system (5) may not always be a feasible component of (4). A simpleheuristic strategy is to divide the search space into several rectangle regions and

Page 10: A novel three-phase trajectory informed search methodology for

70 J Glob Optim (2007) 38:61–77

feasible components

plane: z=f(x*)

attractor

attractor

terminal repeller trajectory

trajectory of the proposed method

x*

••

(a)

(b)

Fig. 4 a shows a landscape of an 2D-objective function. The regions below the planez = f (x� ) fora local minimum x� represent feasible components of (4). b shows a phase portrait of system (5)associated withfx� . Each feasible component in (a) is represented by an attractor of system (5). Themarks ‘€’ represent some escaping points generated by the repelling step and thesolid lines representthe trajectories of system (5) starting from the points marked by ‘ €’. These trajectories are attracted bythe attractors of system (5), most of which converge to the feasible components of (4). The dotted linesrepresent the terminal repelled trajectory of system (6), most of which bypass feasible componentsof (4)

Page 11: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 71

to choose a random point in a region that does not contain previously obtainedlocal minima listed in V. Another strategy, we will call a double repelling step, isas follows: Choose an initial straight rayr(λ), emanated fromr(0) = xs and followthe ray r(λ), with increasing λ, (or employ a mixed cubic and quadratic interpo-lation line search algorithm) until it reaches its 2nd (or ith, i � 2) local minimum(or it hits the boundary of search space) off (r(·)). (The objective function f wouldincrease, pass a �rst local maximum, and reach a �rst local minimum along theray, then pass a second local maximum, and reach a second local minimum andso on.)

4 Simulation results and discussion

4.1 Simulation Benchmark problems

The algorithm described in the previous section has been tested on several benchmarkproblems we have found in the literature ([2, 6, 7, 16]). The descriptions of the testexamples are given in Appendix-B.

Tables1and2show the performance of our proposed algorithms compared to otherglobal optimization methods for low-dimensional problems and high-dimensionalproblems, respectively. The following abbreviations are also used: PRS is the purerandom search; MS the multistart; SA the simulated annealing method of ([22]); EAdenotes the evolution algorithms of ([32]); GA refers the genetic algorithms of ([4,16, 25]); MLSL is the multiple level single linkage method of ([20]); IA is the intervalarithmetic technique of ([28]); TS refers to the tabu search methods of ([3, 6, 10]);MCS is the multilevel coordinate search method of ([18]); TN refers to the tunnelingmethods such as the TRUST algorithm in ([2, 7]). The criterion for comparison isthe number of function evaluations. The results of some compared algorithms (PRS,MS, EA, MLSL, IA, TN in Table 1) are taken from ([2, 7]). The simulation resultsdemonstrate that the proposed method works successfully in producing high-qualitysolutions (some of them are bene�ted from employed hybrid local search algorithmsin Phase I) and is competitive with these state-of-art methods.

Table 1 Number of function evaluations required by different methods to reach a global minimumof low-dimensional test functions

CA GP SH BR H3,4

PRS – 5,125 6,700 4,850 5,280MS – 4,400 – 1,600 2,500EA – 460 – 430 –MLSL – 148 – 206 197IA 326 – 7,424 1,354 –GA 176 191 742 173 201SA 1,903 – 780 505 1,459TS 246 230 274 212 438MCS – 81 141 45 224TN 31 103 72 55 58Proposed 38 199 67 26 22

Page 12: A novel three-phase trajectory informed search methodology for

72 J Glob Optim (2007) 38:61–77

Table 2 Number of function evaluations required by different methods to reach a global minimumof high-dimensional test functions

S4,5 S4,7 S4,10 H6,4 RT10 R10 RT20 R20 DP25 L30

GA 1,086 1,087 1,068 973 1,2078 6,340 – – – –TS 841 833 840 – 5,107 7,062 24,485 34,139 39,374 65,367SA – – – – 1,08,243 1,05,137 81,454 74,512 – 1,06,798MCS 298 178 384 146 1,963 1,233 5,957 4,247 30,376 1,661TN 298 533 469 963 – – – – – –Proposed 443 480 328 65 343 723 1,045 1,743 1,863 324

4.2 Large-scale simulation

To evaluate the performance of the proposed method for a large-scale practical appli-cation, we have performed a simulation with a sonar data set. This is the data setused by Gorman and Sejnowski [13] in their study of the classi�cation of sonar signalsusing a neural network. The data consists of sonar returns collected from two sources:a metal cylinder and a similarly shaped rock. Both objects were lying on a sand surface,and the sonar chirp projected at them from different angles produced the variationin the data. The data contains 111 patterns obtained from the metal cylinder, and 97patterns obtained from the rock. Each pattern is a set of 60 continuous numbers in therange 0.0 – 1.0. For the simulation, a three layer network of size 60-5-1 is considered.This one consists of an input layer, a hidden layer, and an output layer, interconnectedby modi�able weights, represented by links between layers. For each inputx(n) � � 60,n = 1,. . . , 208, the output of this neural network model is given by

zk(n) = φ

⎛⎝

5∑j= 1

wj+ 1φ

⎛⎝

60∑i= 1

w5i+ j+ 6xi(n) + wj+ 6

⎞⎠ + w1

⎞⎠ ,

where w = (w1, . . . , w311)T is a synaptic weight vector andφ: � � � is a nonlinear

activation function (e.g. φ(s) = 1/(1+ exp(Šay)) for a positive gain a). The supervisedtraining of a neural network is then viewed as an unconstrained minimization problemas follows:

minw

f (w),

where the objective function f: � 311 � � is a highly nonlinear mean squared trainingerror (MSE) function.

Figure 5 shows the performance of our proposed algorithm compared to errorback-propagation algorithm (EBP) [15], TRUST algorithm or its variant (TR), geneticalgorithm (GA) and simulated annealing algorithm (SA). The result demonstrates asigni�cant performance improvement of the proposed method compared with otheralgorithms.

5 Concluding remarks

In this paper, a new deterministic method for global optimization has been developed.The proposed method consists of three-phases: Phase I is a typical continuous local

Page 13: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 73

0 100 200 300 400 500 600 700 800

10–1

100

Number of epochs

obje

ctiv

e va

lue

f

SA

Proposed

TR

GA

EBP

Fig. 5 Convergence curve for the sonar problem

search and looks for a local minimum. Phase II is a discrete sup-local search and looksfor a sup-local minimum with the lowest objective value among the neighboring localminima. Phase III is an attractor-based global search and looks for a feasible solutionwith a lower objective value than those values obtained in Phase II.

The method has several features: �rst it does not require a good initial guess(although such a guess would of course reduce the number of steps needed). Second,it is better than stochastic methods such as multi-start methods since it can avoidmany unnecessary efforts in re-determining already known minima. Third, it is verygeneral and could be extended to any global optimization problem. An application ofthe method to more large-scale real-world problems remains to be investigated.

Appendix

Proof of Theorem 1

Proof Let C be a feasible connected component of the setS(x� ) and x � C̄ where C̄is the closure of C. Since

fx� (x) = f (x) Š f (x� ) � 0 or ρ(fx� (x)) = 0

Page 14: A novel three-phase trajectory informed search methodology for

74 J Glob Optim (2007) 38:61–77

the Jacobian of F(x) � Š ρ(fx� (x)) f (x) at x is

JF(x) = Š ρ(fx� (x)) 2f (x) Š f (x) f (x)Tρ (fx� (x))

= Š f (x) f (x)Tρ (fx� (x)).

If x � C, then ρ (fx� (x)) = 0 and so we haveJF(x) = 0. Next if x � ∂C, thenρ (fx� (x)) > 0. Here we assume that∂C is an (n Š 1)-dimensional manifold, which isa generic property by Morse–Sard Theorem. To show that the setC̄ is an attractorof (5), we will show that vTJF(x)v = 0 for all v � Tx(C) and vTJF(x)v < 0 for allv � Tx(C)� where Tx(C) is an (n Š 1)-dimensional tangent space ofC at x and can bedescribed by

Tx(C) = f (x)� = { v : f (x)Tv = 0}.

(1) For any vector v � f (x)� = { v : f (x)Tv = 0}, we havevTJF(x)v = 0.(2) For any vector v �� f (x)� , i.e., v = c f (x) where c is a scalar, we have

vTJF(x)v = Š ρ (fx� (x))� f (x)Tv� 2 < 0

Therefore, C is an attractor of system (5). ��

Test functions

(CA) or (HM) Hump Function (or Six-hump Camelback 2-D FunctionDe�nition: HM(x) = 1.0316285+ 4x2

1 Š 2.1x41 + 1

3x61 + x1x2 Š 4x2

2 + 4x42

Search space:Š5 � xi � 5, i = 1, 2Global minimum: x� = (0.0898,Š0.7126), (Š0.0898, 0.7126); HM(x� ) = 0

(GP) Goldstein and Price FunctionDe�nition: GP(x) =

(1+ (x1 + x2 + 1)2(19Š 14x1 + 13x2

1 Š 14x2 + 6x1x2 + 3x22)

)(30+

((2x1 Š 3x2)2(18Š 32x1 + 12x2

1 Š 48x2 Š 36x1x2 + 27x22)

)Search space:Š2 � xi � 2, i = 1, 2Global minimum: x� = (0,Š1); GP(x� ) = 3

(SH) Shubert Function

De�nition: SH(x) =( ∑5

i= 1 i cos((i + 1)x1 + i

))(∑5i= 1 i cos

((i + 1)x2 + i

))

Search space:Š10 � xi � 10, i = 1, 2Global minimum: 18 global minima and SH(x� ) = Š 186.7309

(BR) Branin RCOS FunctionDe�nition: RC(x) = (x2 Š 5

4π2 x21 + 5

πx1 Š 6)2 + 10(1 Š 1

8π) cos(x1) + 10

Search space:Š5 � x1 � 10, 0� x2 � 15Global minimum: x� = (Šπ , 12.275), (π , 2.275), (9.42478, 2.475); RC(x� ) = 0.397887

(H 3,4) Hartmann FunctionDe�nition: H3,4(x) = Š

∑4i= 1 αi exp

[ ∑3j= 1 Aij(xj Š Pij)

2],α = [ 1, 1.2, 3, 3.2]T

A =

⎡⎢⎢⎣

3.0 10 300.1 10 353.0 10 300.1 10 35

⎤⎥⎥⎦ , P = 10Š4

⎡⎢⎢⎣

6890 1170 26734699 4387 74701091 8732 5547381 5743 8828

⎤⎥⎥⎦ .

Page 15: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 75

Search space: 0� xi � 1, i = 1, 2, 3Global minimum: x� = (0.114614, 0.555649, 0.852547); H3,4(x� ) = Š 3.86278

(H 6,4) Hartmann FunctionDe�nition: H6,4(x) = Š

∑4i= 1 αi exp

[ ∑6j= 1 Bij(xj Š Qij)

2],α = [ 1, 1.2, 3, 3.2]T

B =

⎡⎢⎢⎣

10 3 17 3.05 1.7 80.05 10 17 0.1 8 14

3 3.5 1.7 10 17 817 8 0.05 10 0.1 14

⎤⎥⎥⎦ ,

Q = 10Š4

⎡⎢⎢⎣

1312 1696 5569 124 8283 58862329 4135 8307 3736 1004 99912348 1451 3522 2883 3047 66504047 8828 8732 5743 1091 381

⎤⎥⎥⎦ .

Search space: 0� xi � 1, i = 1,. . . , 6Global minimum: x� = (0.201690, 0.150011, 0.476874, 0.275332, 0.311652, 0.657300);H3,4(x� ) = Š 3.32237

(S4,m) Shekel Function

De�nition: S4,m(x) = Š∑m

j= 1[∑4

i= 1(xi Š Cij)2 + βj

]Š1,β = 110[1, 2, 2, 4, 4, 6, 3, 7, 5, 5]T

C =

⎡⎢⎢⎣

4.0 1.0 8.0 6.0 3.0 2.0 5.0 8.0 6.0 7.04.0 1.0 8.0 6.0 7.0 9.0 5.0 1.0 2.0 3.64.0 1.0 8.0 6.0 3.0 2.0 3.0 8.0 6.0 7.04.0 1.0 8.0 6.0 7.0 9.0 3.0 1.0 2.0 3.6

⎤⎥⎥⎦

Search space: 0� xi � 10, i = 1,. . . , 4Global minimum: x� = (4, 4, 4, 4); S4,5(x� ) = Š 10.1532,S4,7(x� ) = Š 10.4029,S4,10(x� ) = Š 10.5364

(RT n) Rastrigin FunctionDe�nition: RTn(x) = 10n +

∑ni= 1(x

2i Š 10cos(2πxi))

Search space:Š2.56� xi � 5.12, i = 1,. . . , nGlobal minimum: x� = (0,. . . , 0); RTn(x� ) = 0

(Rn) Rosenbrock FunctionDe�nition: Rn(x) =

∑nŠ1i= 1

[100(x2

i Š xi+ 1)2 + (xi Š 1)2

]Search space:Š5 � xi � 10, i = 1, 2,. . . , nGlobal minimum: x� = (1,. . . , 1); Rn(x� ) = 0

(DPn) Dixon & Price FunctionDe�nition: DPn(x) = (x1 Š 1)2 +

∑ni= 2 i(2x2

i Š xiŠ1)2

Search space:Š10 � xi � 10, i = 1,. . . , n

Global minimum: x�i = 2

Š(

2iŠ2

2i

), i = 1,. . . , n; DPn(x� ) = 0

(L n) Levy Function

De�nition: Ln(x) = sin2(πy1) +∑nŠ1

i= 1

[(yi Š 1)2

(1+ 10sin2(πyi + 1)

)]+ (yn Š 1)2

(1+

Page 16: A novel three-phase trajectory informed search methodology for

76 J Glob Optim (2007) 38:61–77

10sin2(2πyn))

yi = 1 + xiŠ14 , i = 1,. . . , n

Search space:Š10 � xi � 10, i = 1,. . . , nGlobal minimum: x� = (1,. . . , 1); Ln(x� ) = 0

Acknowledgements This work was supported by the KOSEF under the grant numberR01-2005-000-10746-0.

References

1. Al Baali, M.: Improved Hessian approximation for the limited memory BFGS method. Numer.Algorithms, 22, 99–112 (1999)

2. Barhen, J., Protopopescu, V., Reister, D.: TRUST: a deterministic algorithm for global optimiza-tion. Science,276, 1094–1097 (1997)

3. Battiti, R., Tecchiolli, G.: The continuous reactive tabu search: blending combinatorial optimiza-tion and stochastic search for global optimization. Ann. Oper. Res.63, 153–188 (1996)

4. Bessaou, M., Siarry, P.: A genetic algorithm with real-value coding to optimize multimodal con-tinuous functions. Struct. Multidisc. Optim. 23, 63–74 (2001)

5. Brown, A.A., Bartholomew-Biggs, M.C.: Some effective methods for unconstrained optimizationbased on the solution of systems of ordinary differential equations. J. Optim. Theory Appl. 62,211–224 (1989)

6. Chelouah, R., Siarry, P.: Tabu search applied to global optimization. Eur. J. Oper. Res.123, 256–270(2000)

7. Cetin, B.C., Barhen, J., Burdick, J.W.: Terminal repeller unconstrained subenergy tunnelling. J.Optim. Theory Appl. 77(1), 97–126 (1993)

8. Chua, L.O., Lin, G.N.: Nonlinear optimization with constraints. IEEE Trans. Circ. Syst. CAS-31,182–188 (1984)

9. Chua, L.O.: Global optimization: a naive approach. IEEE Trans. Circ. Syst. 37, 966–969 (1990)10. Cvijovic, D., Klinowski, J.: Taboo search: An approach to the multiple minima problem. Science

267, 664 (1995)11. Di Fiore, C., Fanelli, S., Lepore, F., Zellini, P.: Matrix algebras in Quasi-Newton methods for

unconstrained minimization. Numer. Math. 94, 479–500 (2003)12. Floudas, C.A., Pardalos, P.M.: Recent advances in Global Optimization. Princeton series in Com-

puter science, NJ (1992)13. Gorman, R.P., Sejnowski, T.J.: Analysis of hidden units in a layered network trained to classify

sonar targets. Neural Netw.1, 75–89 (1988)14. Guckenheimer, J., Homes, P.: Nonlinear Oscillations, Dynamical Systems, and Bifurcations of

Vector Fields. Applied math. sci. vol. 42, Springer, New York (1986)15. Haykin, S.: Neural Networks: A Comprehensive Foundation. Prentice-Hall, New York (1999)16. Hedar, A., Fukushima, M.: Minimizing multimodal functions by simplex coding genetic algorithm.

Optim. Methods Softw. 18, 265–282 (2003)17. Horst, R., Pardalos, P.M.: Handbook of Global Optimization, Kluwer Academic Publishers,

Dordrecht (1994)18. Huyer, W., Neumaier, A.: Global optimization by multilevel coordinate search. J. Global Optim.

14, 331–355 (1999)19. Jongen, H.Th., Jonker, P., Twilt, F.:Nonlinear Optimization in � n, Peter Lang Verlag, Frankfurt

(1986)20. Kan, A.H.G.R., Timmer, G.T.: A stochastic approach to global optimization. In: Boggs, P.T., Byrd,

R.H., Schnabel, R.B. (eds.), Numerical Optimization pp. 245–262. SIAM, Philadelphia, PA (1985)21. Kelley, C.T.: Detection and remediation of stagnation in the Nelder-Mead algorithm using a

suf�cient decrease condition. SIAM J. Optim. 10, 43–55 (1999)22. Kirkpatrick, S., Gelatt, C.D., Timer, G.T.: Optimization by simulated annealing. Science 220,

671–680 (1983)23. Lee, J.: Dynamic gradient approaches to compute the closest unstable equilibrium point for

stability region estimate and their computational limitations. IEEE Trans. Autom. Control 48(2),321–324 (2003)

24. Levy, A.V., Montalvo, A.: The tunnelling algorithm for the global minimization of functions.SIAM J. Sci. Stati. Comput. 6, 15–29 (1985)

Page 17: A novel three-phase trajectory informed search methodology for

J Glob Optim (2007) 38:61–77 77

25. Michalewicz, Z.: Genetic Algorithms + Data Structures = Evolution Programs. Springer, Berlin(1996)

26. Nelder, J.A., Mead, R.: A simplex method for function minimization. Comput. J. 7, 308–313 (1965)27. Nocedal, J.N., Wright, S.J.:Numerical Optimization. Springer-verlag, New York, Heidelberg,

Berlin (1999)28. Ratschek, H., Rokne, J.: New Computer Methods for Global Optimization. Ellis Horwood,

Chichester, UK (1988)29. Tank, D.W., Hop�eld, J.J.: Simple ‘neural’ optimization networks: an A/D converter, signal deci-

sion circuit, and a linear porgramming circuit. IEEE Trans. Circ. Sys. CAS-33(5), 533–541 (1986)30. Vavasis, S.:Nonlinear Optimization: Complexity Issues. Springer-Verlag, New York (1990)31. Yao, Y.: Dynamic tunnelling algorithm for global optimization. IEEE Trans. Sys. Man Cybern. 19,

1222–1230 (1989)32. Yong, L., Lishan, K., Evans, D.J.: The annealing evolution algorithm as function optimizer. Parallel

Comput. 21, 389 (1995)