Sie sind auf Seite 1von 41

Metaheuristics in Combinatorial Optimization: Overview

and Conceptual Comparison


CHRISTIAN BLUM
Université Libre de Bruxelles

AND

ANDREA ROLI
Università degli Studi di Bologna

The field of metaheuristics for the application to combinatorial optimization problems is


a rapidly growing field of research. This is due to the importance of combinatorial
optimization problems for the scientific as well as the industrial world. We give a survey
of the nowadays most important metaheuristics from a conceptual point of view. We
outline the different components and concepts that are used in the different
metaheuristics in order to analyze their similarities and differences. Two very
important concepts in metaheuristics are intensification and diversification. These are
the two forces that largely determine the behavior of a metaheuristic. They are in some
way contrary but also complementary to each other. We introduce a framework, that we
call the I&D frame, in order to put different intensification and diversification
components into relation with each other. Outlining the advantages and disadvantages
of different metaheuristic approaches we conclude by pointing out the importance of
hybridization of metaheuristics as well as the integration of metaheuristics and other
methods for optimization.

Categories and Subject Descriptors: G.2.1 [Discrete Mathematics]: Combinatorics—


combinatorial algorithms; I.2.8 [Artificial Intelligence]: Problem Solving, Control
Methods, and Search—heuristic methods
General Terms: Algorithms
Additional Key Words and Phrases: Metaheuristics, combinatorial optimization,
intensification, diversification.

C. Blum acknowledges support by the “Metaheuristics Network,” a Research Training Network funded by
the Improving Human Potential program of the CEC, contract HPRN-CT-1999-00106.
A. Roli acknowledges support by the CEC through a “Marie Curie Training Site” fellowship, contract HPMT-
CT-2000-00032.
The information provided is the sole responsibility of the authors and does not reflect the Community’s opin-
ion. The Community is not responsible for any use that might be made of data appearing in this publication.
Authors’ addresses: C. Blum, Université Libre de Bruxelles, IRIDIA, Avenue Franklin Roosevelt 50, CP
194/6, 1050 Brussels, Belgium; email: cblum@ulb.ac.be; A. Roli, DEIA—Università degli Studi di Bologna,
Viale Risorgimento, 2-Bologna, Italy; email: aroli@deis.unibo.it.
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted
without fee provided that copies are not made or distributed for profit or direct commercial advantage and
that copies show this notice on the first page or initial screen of a display along with the full citation. Copy-
rights for components of this work owned by others than ACM must be honored. Abstracting with credit is
permitted. To copy otherwise, to republish, to post on servers, to redistribute to lists, or to use any component
of this work in other works requires prior specific permission and/or a fee. Permissions may be requested
from Publications Dept., ACM, Inc., 1515 Broadway, New York, NY 10036 USA, fax: +1 (212) 869-0481, or
permissions@acm.org.
°2003
c ACM 0360-0300/03/0900-0268 $5.00

ACM Computing Surveys, Vol. 35, No. 3, September 2003, pp. 268–308.
Metaheuristics in Combinatorial Optimization 269

1. INTRODUCTION problems, many algorithms to tackle them


have been developed. These algorithms
Many optimization problems of practical
can be classified as either complete or
as well as theoretical importance con-
approximate algorithms. Complete algo-
sist of the search for a “best” configu-
rithms are guaranteed to find for every
ration of a set of variables to achieve
finite size instance of a CO problem an
some goals. They seem to divide naturally
optimal solution in bounded time (see
into two categories: those where solutions
Papadimitriou and Steiglitz [1982] and
are encoded with real-valued variables,
Nemhauser and Wolsey [1988]). Yet, for
and those where solutions are encoded
CO problems that are N P-hard [Garey
with discrete variables. Among the lat-
and Johnson 1979], no polynomial time
ter ones we find a class of problems
algorithm exists, assuming that P 6= N P.
called Combinatorial Optimization (CO)
Therefore, complete methods might need
problems. According to Papadimitriou and
exponential computation time in the
Steiglitz [1982], in CO problems, we are
worst-case. This often leads to computa-
looking for an object from a finite—or pos-
tion times too high for practical purposes.
sibly countably infinite—set. This object
Thus, the use of approximate methods to
is typically an integer number, a subset, a
solve CO problems has received more and
permutation, or a graph structure.
more attention in the last 30 years. In ap-
Definition 1.1. A Combinatorial Opti- proximate methods we sacrifice the guar-
mization problem P = (S, f ) can be de- antee of finding optimal solutions for the
fined by: sake of getting good solutions in a signifi-
cantly reduced amount of time.
—a set of variables X = {x1 , . . . , xn };
Among the basic approximate meth-
—variable domains D1 , . . . , Dn ; ods we usually distinguish between con-
—constraints among variables; structive methods and local search meth-
—an objective function f to be minimi- ods. Constructive algorithms generate
zed,1 where f : D1 × · · · × Dn → IR + ; solutions from scratch by adding—to
an initially empty partial solution—
The set of all possible feasible assignments components, until a solution is complete.
is They are typically the fastest approximate
S = {s = {(x1 , v1 ), . . . , (xn , vn )} | vi ∈ Di , s methods, yet they often return solutions
of inferior quality when compared to lo-
satisfies all the constraints}. cal search algorithms. Local search algo-
rithms start from some initial solution and
S is usually called a search (or solution) iteratively try to replace the current solu-
space, as each element of the set can be tion by a better solution in an appropri-
seen as a candidate solution. To solve a ately defined neighborhood of the current
combinatorial optimization problem one solution, where the neighborhood is for-
has to find a solution s∗ ∈ S with minimum mally defined as follows:
objective function value, that is, f (s∗ ) ≤
f (s) ∀s ∈ S. s∗ is called a globally optimal Definition 1.2. A neighborhood struc-
solution of (S, f ) and the set S ∗ ⊆ S is ture is a function N : S → 2S that assigns
called the set of globally optimal solutions. to every s ∈ S a set of neighbors N (s) ⊆ S.
N (s) is called the neighborhood of s.
Examples for CO problems are the
Travelling Salesman problem (TSP), the The introduction of a neighborhood
Quadratic Assignment problem (QAP), structure enables us to define the concept
Timetabling and Scheduling problems. of locally minimal solutions.
Due to the practical importance of CO
Definition 1.3. A locally minimal so-
1 As maximizing an objective function f is the same
lution (or local minimum) with respect to
as minimizing − f , in this work we will deal, without a neighborhood structure N is a solution
loss of generality, with minimization problems. ŝ such that ∀ s ∈ N (ŝ) : f (ŝ) ≤ f (s). We

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


270 C. Blum and A. Roli

call ŝ a strict locally minimal solution if heuristics may be high (or low) level pro-
f (ŝ) < f (s) ∀ s ∈ N (ŝ). cedures, or a simple local search, or just a
construction method.” [Voß et al. 1999].
In the last 20 years, a new kind “Metaheuristics are typically high-level
of approximate algorithm has emerged strategies which guide an underlying,
which basically tries to combine basic more problem specific heuristic, to in-
heuristic methods in higher level frame- crease their performance. The main goal
works aimed at efficiently and effec- is to avoid the disadvantages of iterative
tively exploring a search space. These improvement and, in particular, multiple
methods are nowadays commonly called descent by allowing the local search to es-
metaheuristics.2 The term metaheuristic, cape from local optima. This is achieved by
first introduced in Glover [1986], derives
either allowing worsening moves or gener-
from the composition of two Greek words.
ating new starting solutions for the local
Heuristic derives from the verb heuriskein
search in a more “intelligent” way than
(²υρισ κ²ιν) which means “to find”, while
just providing random initial solutions.
the suffix meta means “beyond, in an up-
Many of the methods can be interpreted
per level”. Before this term was widely
as introducing a bias such that high qual-
adopted, metaheuristics were often called
ity solutions are produced quickly. This
modern heuristics [Reeves 1993].
bias can be of various forms and can be
This class of algorithms includes3 —but
cast as descent bias (based on the ob-
is not restricted to—Ant Colony Opti-
jective function), memory bias (based on
mization (ACO), Evolutionary Computa-
previously made decisions) or experience
tion (EC) including Genetic Algorithms
bias (based on prior performance). Many
(GA), Iterated Local Search (ILS), Sim-
of the metaheuristic approaches rely on
ulated Annealing (SA), and Tabu Search
probabilistic decisions made during the
(TS). Up to now there is no commonly ac-
search. But, the main difference to pure
cepted definition for the term metaheuris-
random search is that in metaheuris-
tic. It is just in the last few years that some
tic algorithms randomness is not used
researchers in the field tried to propose a
blindly but in an intelligent, biased form.”
definition. In the following we quote some
[Stützle 1999b].
of them:
“A metaheuristic is a set of concepts that
“A metaheuristic is formally defined
can be used to define heuristic methods
as an iterative generation process which
that can be applied to a wide set of dif-
guides a subordinate heuristic by combin-
ferent problems. In other words, a meta-
ing intelligently different concepts for ex-
heuristic can be seen as a general algo-
ploring and exploiting the search space,
rithmic framework which can be applied to
learning strategies are used to struc-
different optimization problems with rel-
ture information in order to find effici-
atively few modifications to make them
ently near-optimal solutions.” [Osman and
adapted to a specific problem.” [Meta-
Laporte 1996].
heuristics Network Website 2000].
“A metaheuristic is an iterative master
Summarizing, we outline fundamen-
process that guides and modifies the op-
tal properties which characterize meta-
erations of subordinate heuristics to effi-
heuristics:
ciently produce high-quality solutions. It
may manipulate a complete (or incom- —Metaheuristics are strategies that
plete) single solution or a collection of so- “guide” the search process.
lutions at each iteration. The subordinate —The goal is to efficiently explore the
search space in order to find (near-)
2 The increasing importance of metaheuristics is un- optimal solutions.
derlined by the biannual Metaheuristics Interna-
tional Conference (MIC). The 5th is being held in
—Techniques which constitute meta-
Kyoto in August 2003 (http://www-or.amp.i.kyoto- heuristic algorithms range from sim-
u.ac.jp/mic2003/). ple local search procedures to complex
3 In alphabetical order. learning processes.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 271

—Metaheuristic algorithms are approxi- plored or which do not provide high quality
mate and usually non-deterministic. solutions.
—They may incorporate mechanisms to The search strategies of different meta-
avoid getting trapped in confined areas heuristics are highly dependent on the
of the search space. philosophy of the metaheuristic itself.
Comparing the strategies used in differ-
—The basic concepts of metaheuristics
ent metaheuristics is one of the goals
permit an abstract level description.
of Section 5. There are several differ-
—Metaheuristics are not problem-specific. ent philosophies apparent in the existing
—Metaheuristics may make use of metaheuristics. Some of them can be seen
domain-specific knowledge in the form as “intelligent” extensions of local search
of heuristics that are controlled by the algorithms. The goal of this kind of meta-
upper level strategy. heuristic is to escape from local minima
—Todays more advanced metaheuristics in order to proceed in the exploration of
use search experience (embodied in the search space and to move on to find
some form of memory) to guide the other hopefully better local minima. This
search. is for example the case in Tabu Search,
Iterated Local Search, Variable Neighbor-
In short, we could say that metaheuris- hood Search, GRASP and Simulated An-
tics are high level strategies for explor- nealing. These metaheuristics (also called
ing search spaces by using different meth- trajectory methods) work on one or sev-
ods. Of great importance hereby is that a eral neighborhood structure(s) imposed on
dynamic balance is given between diversi- the members (the solutions) of the search
fication and intensification. The term di- space.
versification generally refers to the explo- We can find a different philosophy in
ration of the search space, whereas the algorithms like Ant Colony Optimization
term intensification refers to the exploita- and Evolutionary Computation. They in-
tion of the accumulated search experience. corporate a learning component in the
These terms stem from the Tabu Search sense that they implicitly or explicitly
field [Glover and Laguna 1997] and it is try to learn correlations between deci-
important to clarify that the terms ex- sion variables to identify high quality ar-
ploration and exploitation are sometimes eas in the search space. This kind of
used instead, for example in the Evo- metaheuristic performs, in a sense, a bi-
lutionary Computation field [Eiben and ased sampling of the search space. For in-
Schippers 1998], with a more restricted stance, in Evolutionary Computation this
meaning. In fact, the notions of exploita- is achieved by recombination of solutions
tion and exploration often refer to rather and in Ant Colony Optimization by sam-
short-term strategies tied to randomness, pling the search space in every iteration
whereas intensification and diversifica- according to a probability distribution.
tion also refer to medium- and long-term The structure of this work is as follows:
strategies based on the usage of mem- There are several approaches to classify
ory. The use of the terms diversification metaheuristics according to their proper-
and intensification in their initial mean- ties. In Section 2, we briefly list and sum-
ing becomes more and more accepted by marize different classification approaches.
the whole field of metaheuristics. There- Section 3 and Section 4 are devoted to a
fore, we use them throughout the article. description of the most important meta-
The balance between diversification and heuristics nowadays. Section 3 describes
intensification as mentioned above is im- the most relevant trajectory methods and,
portant, on one side to quickly identify re- in Section 4, we outline population-based
gions in the search space with high qual- methods. Section 5 aims at giving a unify-
ity solutions and on the other side not ing view on metaheuristics with respect
to waste too much time in regions of the to the way they achieve intensification
search space which are either already ex- and diversification. This is done by the

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


272 C. Blum and A. Roli

introduction of a unifying framework, the the use of memory in Tabu Search is not
I&D frame. Finally, Section 6 offers some nature-inspired as well.
conclusions and an outlook to the future. Population-based vs. single point search.
We believe that it is hardly possible to Another characteristic that can be used
produce a completely accurate survey of for the classification of metaheuristics is
metaheuristics that is doing justice to ev- the number of solutions used at the same
ery viewpoint. Moreover, a survey of an time: Does the algorithm work on a popu-
immense area such as metaheuristics has lation or on a single solution at any time?
to focus on certain aspects and therefore Algorithms working on single solutions
has unfortunately to neglect other aspects. are called trajectory methods and encom-
Therefore, we want to clarify at this point pass local search-based metaheuristics,
that this survey is done from the concep- like Tabu Search, Iterated Local Search
tual point of view. We want to outline the and Variable Neighborhood Search. They
different concepts that are used in differ- all share the property of describing a tra-
ent metaheuristics in order to analyze the jectory in the search space during the
similarities and the differences between search process. Population-based meta-
them. We do not go into the implementa- heuristics, on the contrary, perform search
tion of metaheuristics, which is certainly processes which describe the evolution of
an important aspect of metaheuristics a set of points in the search space.
research with respect to the increas- Dynamic vs. static objective function.
ing importance of efficiency and software Metaheuristics can also be classified ac-
reusability. We refer the interested reader cording to the way they make use of the
to Whitley [1989], Grefenstette [1990], objective function. While some algorithms
Fink and Voß [1999], Schaerf et al. [2000], keep the objective function given in the
and Voß and Woodruff [2002]. problem representation “as it is”, some
others, like Guided Local Search (GLS),
modify it during the search. The idea be-
2. CLASSIFICATION OF METAHEURISTICS
hind this approach is to escape from lo-
There are different ways to classify and cal minima by modifying the search land-
describe metaheuristic algorithms. De- scape. Accordingly, during the search the
pending on the characteristics selected objective function is altered by trying to in-
to differentiate among them, several corporate information collected during the
classifications are possible, each of them search process.
being the result of a specific viewpoint. One vs. various neighborhood structures.
We briefly summarize the most important Most metaheuristic algorithms work on
ways of classifying metaheuristics. one single neighborhood structure. In
other words, the fitness landscape topol-
Nature-inspired vs. non-nature in- ogy does not change in the course of the
spired. Perhaps, the most intuitive way algorithm. Other metaheuristics, such as
of classifying metaheuristics is based on Variable Neighborhood Search (VNS), use
the origins of the algorithm. There are a set of neighborhood structures which
nature-inspired algorithms, like Genetic gives the possibility to diversify the search
Algorithms and Ant Algorithms, and by swapping between different fitness
non nature-inspired ones such as Tabu landscapes.
Search and Iterated Local Search. In Memory usage vs. memory-less methods.
our opinion this classification is not very A very important feature to classify meta-
meaningful for the following two reasons. heuristics is the use they make of the
First, many recent hybrid algorithms do search history, that is, whether they use
not fit either class (or, in a sense, they memory or not.4 Memory-less algorithms
fit both at the same time). Second, it is
sometimes difficult to clearly attribute an 4 Here we refer to the use of adaptive memory, in con-
algorithm to one of the two classes. So, trast to rather rigid memory, as used for instance in
for example, one might ask the question if Branch & Bound.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 273

perform a Markov process, as the infor- s ← GenerateInitialSolution()


mation they exclusively use to determine repeat
the next action is the current state of the s ← Improve(N (s))
search process. There are several different until no improvement is possible
ways of making use of memory. Usually we Fig. 1. Algorithm: Iterative
differentiate between the use of short term Improvement.
and long term memory. The first usually
keeps track of recently performed moves, trajectories which can not be subdivided
visited solutions or, in general, decisions in those two phases. The characteristics of
taken. The second is usually an accumu- the trajectory provide information about
lation of synthetic parameters about the the behavior of the algorithm and its effec-
search. The use of memory is nowadays tiveness with respect to the instance that
recognized as one of the fundamental ele- is tackled. It is worth underlining that
ments of a powerful metaheuristic. the dynamics is the result of the combina-
In the following, we describe the most tion of algorithm, problem representation
important metaheuristics according to and problem instance. In fact, the problem
the single point vs. population-based representation together with the neigh-
search classification, which divides meta- borhood structures define the search land-
heuristics into trajectory methods and scape; the algorithm describes the strat-
population-based methods. This choice is egy used to explore the landscape and,
motivated by the fact that this categoriza- finally, the actual search space character-
tion permits a clearer description of the al- istics are defined by the problem instance
gorithms. Moreover, a current trend is the to be solved.
hybridization of methods in the direction We will first describe basic local search
of the integration of single point search al- algorithms, before we proceed with the
gorithms in population-based ones. In the survey of more complex strategies. Fi-
following two sections, we give a detailed nally, we deal with algorithms that are
description of nowadays most important general explorative strategies which may
metaheuristics. incorporate other trajectory methods as
components.
3. TRAJECTORY METHODS
3.1. Basic Local Search: Iterative
In this section we outline metaheuristics Improvement
called trajectory methods. The term trajec-
tory methods is used because the search The basic local search is usually called it-
process performed by these methods is erative improvement, since each move5 is
characterized by a trajectory in the search only performed if the resulting solution is
space. Hereby, a successor solution may or better than the current solution. The algo-
may not belong to the neighborhood of the rithm stops as soon as it finds a local mini-
current solution. mum. The high level algorithm is sketched
The search process of trajectory meth- in Figure 1.
ods can be seen as the evolution in (dis- The function Improve(N (s)) can be in the
crete) time of a discrete dynamical sys- extremes either a first improvement, or
tem [Bar-Yam 1997; Devaney 1989]. The a best improvement function, or any in-
algorithm starts from an initial state (the termediate option. The former scans the
initial solution) and describes a trajectory neighborhood N (s) and chooses the first
in the state space. The system dynamics solution that is better than s, the latter ex-
depends on the strategy used; simple al- haustively explores the neighborhood and
gorithms generate a trajectory composed returns one of the solutions with the low-
of two parts: a transient phase followed est objective function value. Both methods
by an attractor (a fixed point, a cycle or
a complex attractor). Algorithms with ad- 5A move is the choice of a solution s0 from the neigh-
vanced strategies generate more complex borhood N (s) of a solution s.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


274 C. Blum and A. Roli

s ← GenerateInitialSolution() tically constructed) and by initializing the


T ← T0 so-called temperature parameter T . Then,
while termination conditions not met do
s0 ← PickAtRandom(N (s))
at each iteration a solution s0 ∈ N (s) is ran-
if ( f (s0 ) < f (s)) then domly sampled and it is accepted as new
s ← s0 % s0 replaces s current solution depending on f (s), f (s0 )
else and T . s0 replaces s if f (s0 ) < f (s) or, in case
Accept s0 as new solution with probability p(T, s0 , s) f (s0 ) ≥ f (s), with a probability which is a
endif function of T and f (s0 ) − f (s). The proba-
Update(T )
bility is generally computed following the
endwhile 0
Boltzmann distribution exp(− f (s )− T
f (s)
).
Fig. 2. Algorithm: Simulated Annealing (SA). The temperature T is decreased6 during
the search process, thus at the beginning
stop at local minima. Therefore, their per- of the search the probability of accepting
formance strongly depends on the defini- uphill moves is high and it gradually de-
tion of S, f and N . The performance of creases, converging to a simple iterative
iterative improvement procedures on CO improvement algorithm. This process is
problems is usually quite unsatisfactory. analogous to the annealing process of met-
Therefore, several techniques have been als and glass, which assume a low en-
developed to prevent algorithms from get- ergy configuration when cooled with an
ting trapped in local minima, which is appropriate cooling schedule. Regarding
done by adding mechanisms that allow the search process, this means that the
them to escape from local minima. This algorithm is the result of two combined
also implies that the termination con- strategies: random walk and iterative im-
ditions of metaheuristic algorithms are provement. In the first phase of the search,
more complex than simply reaching a lo- the bias toward improvements is low and
cal minimum. Indeed, possible termina- it permits the exploration of the search
tion conditions include: maximum CPU space; this erratic component is slowly de-
time, a maximum number of iterations, creased thus leading the search to con-
a solution s with f (s) less than a prede- verge to a (local) minimum. The probabil-
fined threshold value is found, or the max- ity of accepting uphill moves is controlled
imum number of iterations without im- by two factors: the difference of the ob-
provements is reached. jective functions and the temperature. On
the one hand, at fixed temperature, the
3.2. Simulated Annealing higher the difference f (s0 )− f (s), the lower
the probability to accept a move from s to
Simulated Annealing (SA) is commonly s0 . On the other hand, the higher T , the
said to be the oldest among the meta- higher the probability of uphill moves.
heuristics and surely one of the first al- The choice of an appropriate cooling
gorithms that had an explicit strategy to schedule is crucial for the performance of
escape from local minima. The origins of the algorithm. The cooling schedule de-
the algorithm are in statistical mechan- fines the value of T at each iteration k,
ics (Metropolis algorithm) and it was first Tk+1 = Q(Tk , k), where Q(Tk , k) is a func-
presented as a search algorithm for CO tion of the temperature and of the itera-
problems in Kirkpatrick et al. [1983] and tion number. Theoretical results on non-
Cerny [1985]. The fundamental idea is to homogeneous Markov chains [Aarts et al.
allow moves resulting in solutions of worse 1997] state that under particular condi-
quality than the current solution (uphill tions on the cooling schedule, the algo-
moves) in order to escape from local min- rithm converges in probability to a global
ima. The probability of doing such a move
is decreased during the search. The high
level algorithm is described in Figure 2. 6T is not necessarily decreased in a monotonic fash-
The algorithm starts by generating an ion. Elaborate cooling schemes also incorporate an
initial solution (either randomly or heuris- occasional increase of the temperature.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 275

minimum for k → ∞. More precisely: s ← GenerateInitialSolution()


TabuList ← ∅
∃0 ∈ IR such that while termination conditions not met do
lim p( global minimum found after s ← ChooseBestOf(N (s) \ TabuList)
k→∞ Update(TabuList)
k steps) = 1 endwhile
X∞ µ ¶
0 Fig. 3. Algorithm: Simple Tabu Search (TS).
iff exp =∞
Tk
k=1
a trajectory in the state space in which
A particular cooling schedule that fulfils the successor state is chosen depending
the hypothesis for the convergence is the only on the incumbent one. This means
one that follows a logarithmic law: Tk+1 = that basic SA is memory-less. However,
0
log(k+k0 )
(where k0 is a constant). Unfortu- the use of memory can be beneficial for
nately, cooling schedules which guarantee SA approaches (see, e.g., Chardaire et al.
the convergence to a global optimum are [1995]).
not feasible in applications, because they SA has been applied to several CO prob-
are too slow for practical purposes. There- lems, such as the Quadratic Assignment
fore, faster cooling schedules are adopted Problem (QAP) [Connolly 1990] and the
in applications. One of the most used fol- Job Shop Scheduling (JSS) problem [Van
lows a geometric law: Tk+1 = αTk , where Laarhoven et al. 1992]. References to
α ∈ (0, 1), which corresponds to an expo- other applications can be found in Aarts
nential decay of the temperature. and Lenstra [1997], Ingber [1996] and
The cooling rule may vary during the Fleischer [1995]. SA is nowadays used as a
search, with the aim of tuning the bal- component in metaheuristics, rather than
ance between diversification and intensi- applied as stand-alone search algorithm.
fication. For example, at the beginning of Variants of SA called Threshold Accept-
the search, T might be constant or lin- ing and The Great Deluge Algorithm were
early decreasing, in order to sample the presented by Dueck and Scheuer [1990]
search space; then, T might follow a rule and Dueck [1993].
such as the geometric one, to converge to
a local minimum at the end of the search. 3.3. Tabu Search
More successful variants are nonmono-
tonic cooling schedules (e.g., see Osman Tabu Search (TS) is among the most cited
[1993] and Lundy and Mees [1986]). Non- and used metaheuristics for CO prob-
monotonic cooling schedules are charac- lems. TS basic ideas were first introduced
terized by alternating phases of cooling in Glover [1986], based on earlier ideas for-
and reheating, thus providing an oscillat- mulated in Glover [1977].7 A description of
ing balance between diversification and the method and its concepts can be found
intensification. in Glover and Laguna [1997]. TS explicitly
The cooling schedule and the initial tem- uses the history of the search, both to es-
perature should be adapted to the partic- cape from local minima and to implement
ular problem instance, since the cost of an explorative strategy. We will first de-
escaping from local minima depends on scribe a simple version of TS, to introduce
the structure of the search landscape. A the basic concepts. Then, we will explain a
simple way of empirically determining the more applicable algorithm and finally we
starting temperature T0 is to initially sam- will discuss some improvements.
ple the search space with a random walk The simple TS algorithm (see Figure 3)
to roughly evaluate the average and the applies a best improvement local search
variance of objective function values. But as basic ingredient and uses a short term
also more elaborate schemes can be imple- memory to escape from local minima and
mented [Ingber 1996].
The dynamic process described by SA is 7 Relatedideas were labelled steepest ascent/mildest
a Markov chain [Feller 1968], as it follows descent method in Hansen [1986].

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


276 C. Blum and A. Roli

to avoid cycles. The short term memory while it is decreased if there are no im-
is implemented as a tabu list that keeps provements (thus intensification should
track of the most recently visited solu- be boosted). More advanced ways to cre-
tions and forbids moves toward them. The ate dynamic tabu tenure are described
neighborhood of the current solution is in Glover [1990].
thus restricted to the solutions that do not However, the implementation of short
belong to the tabu list. In the following term memory as a list that contains com-
we will refer to this set as allowed set. plete solutions is not practical, because
At each iteration the best solution from managing a list of solutions is highly in-
the allowed set is chosen as the new cur- efficient. Therefore, instead of the solu-
rent solution. Additionally, this solution is tions themselves, solution attributes are
added to the tabu list and one of the so- stored.10 Attributes are usually compo-
lutions that were already in the tabu list nents of solutions, moves, or differences
is removed (usually in a FIFO order). Due between two solutions. Since more than
to this dynamic restriction of allowed so- one attribute can be considered, a tabu
lutions in a neighborhood, TS can be con- list is introduced for each of them. The set
sidered as a dynamic neighborhood search of attributes and the corresponding tabu
technique [Stützle 1999b]. The algorithm lists define the tabu conditions which are
stops when a termination condition is met. used to filter the neighborhood of a solu-
It might also terminate if the allowed set tion and generate the allowed set. Storing
is empty, that is, if all the solutions in N (s) attributes instead of complete solutions is
are forbidden by the tabu list.8 much more efficient, but it introduces a
The use of a tabu list prevents from loss of information, as forbidding an at-
returning to recently visited solutions, tribute means assigning the tabu status to
therefore it prevents from endless cycling9 probably more than one solution. Thus, it
and forces the search to accept even up- is possible that unvisited solutions of good
hill moves. The length l of the tabu list quality are excluded from the allowed set.
(i.e., the tabu tenure) controls the mem- To overcome this problem, aspiration cri-
ory of the search process. With small tabu teria are defined which allow to include a
tenures the search will concentrate on solution in the allowed set even if it is for-
small areas of the search space. On the bidden by tabu conditions. Aspiration cri-
opposite, a large tabu tenure forces the teria define the aspiration conditions that
search process to explore larger regions, are used to construct the allowed set. The
because it forbids revisiting a higher num- most commonly used aspiration criterion
ber of solutions. The tabu tenure can be selects solutions which are better than the
varied during the search, leading to more current best one. The complete algorithm,
robust algorithms. An example can be as described above, is reported in Figure 4.
found in Taillard [1991], where the tabu Tabu lists are only one of the possible
tenure is periodically reinitialized at ran- ways of taking advantage of the history
dom from the interval [l min , l max ]. A more of the search. They are usually identi-
advanced use of a dynamic tabu tenure is fied with the usage of short term memory.
presented in Battiti and Tecchiolli [1994] Information collected during the whole
and Battiti and Protasi [1997], where search process can also be very useful,
the tabu tenure is increased if there especially for a strategic guidance of the
is evidence for repetitions of solutions algorithm. This kind of long-term mem-
(thus a higher diversification is needed), ory is usually added to TS by referring to
four principles: recency, frequency, quality
and influence. Recency-based memory
8 Strategies for avoiding to stop the search when the records for each solution (or attribute)
allowed set is empty include the choice of the least
recently visited solution, even if it is tabu.
9 Cycles of higher period are possible, since the tabu 10 In addition to storing attributes, some longer term
list has a finite length l which is smaller than the TS strategies also keep complete solutions (e.g., elite
cardinality of the search space. solutions) in the memory.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 277

s ← GenerateInitialSolution()
InitializeTabuLists(TL1 , . . . , TLr )
k←0
while termination conditions not met do
AllowedSet(s, k) ← {s0 ∈ N (s) | s does not violate a tabu condition,
or it satisfies at least one aspiration condition}
s ← ChooseBestOf(AllowedSet(s, k))
UpdateTabuListsAndAspirationConditions()
k ←k+1
endwhile

Fig. 4. Algorithm: Tabu Search (TS).

while termination conditions not met do


s ← ConstructGreedyRandomizedSolution() % see Figure 6
ApplyLocalSearch(s)
MemorizeBestFoundSolution()
endwhile

Fig. 5. Algorithm: Greedy Randomized Adaptive Search Procedure (GRASP).

the most recent iteration it was involved proaches dominate the Job Shop Schedu-
in. Orthogonally, frequency-based mem- ling (JSS) problem area (see, e.g., Nowicki
ory keeps track of how many times each and Smutnicki [1996]) and the Vehicle
solution (attribute) has been visited. This Routing (VR) area [Gendreau et al. 2001].
information identifies the regions (or the Further current applications can be found
subsets) of the solution space where the at [Tabu Search website 2003].
search was confined, or where it stayed for
a high number of iterations. This kind of 3.4. Explorative Local Search Methods
information about the past is usually ex-
ploited to diversify the search. The third In this section, we present more recently
principle (i.e., quality) refers to the ac- proposed trajectory methods. These are
cumulation and extraction of information the Greedy Randomized Adaptive Search
from the search history in order to identify Procedure (GRASP), Variable Neighbor-
good solution components. This informa- hood Search (VNS), Guided Local Search
tion can be usefully integrated in the so- (GLS) and Iterated Local Search (ILS).
lution construction. Other metaheuristics
(e.g., Ant Colony Optimization) explicitly 3.4.1. GRASP. The Greedy Randomized
use this principle to learn about good com- Adaptive Search Procedure (GRASP), see
binations of solution components. Finally, Feo and Resende [1995] and Pitsoulis and
influence is a property regarding choices Resende [2002], is a simple metaheuristic
made during the search and can be used that combines constructive heuristics and
to indicate which choices have shown to be local search. Its structure is sketched in
the most critical. In general, the TS field is Figure 5. GRASP is an iterative procedure,
a rich source of ideas. Many of these ideas composed of two phases: solution construc-
and strategies have been and are currently tion and solution improvement. The best
adopted by other metaheuristics. found solution is returned upon termina-
TS has been applied to most CO prob- tion of the search process.
lems; examples for successful applica- The solution construction mechanism
tions are the Robust Tabu Search to the (see Figure 6) is characterized by two
QAP [Taillard 1991], the Reactive Tabu main ingredients: a dynamic constructive
Search to the MAXSAT problem [Battiti heuristic and randomization. Assuming
and Protasi 1997], and to assignment that a solution s consists of a subset of
problems [Dell’Amico et al. 1999]. TS ap- a set of elements (solution components),

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


278 C. Blum and A. Roli

s←∅ % s denotes a partial solution in this case


α ← DetermineCandidateListLength() % definition of the RCL length
while solution not complete do
RCLα ← GenerateRestrictedCandidateList(s)
x ← SelectElementAtRandom(RCLα )
s ← s ∪ {x}
UpdateGreedyFunction(s) % update of the heuristic values (see text)
endwhile

Fig. 6. Greedy randomized solution construction.

the solution is constructed step-by-step by The second phase of the algorithm is a


adding one new element at a time. The local search process, which may be a ba-
choice of the next element is done by pick- sic local search algorithm such as iterative
ing it uniformly at random from a can- improvement, or a more advanced tech-
didate list. The elements are ranked by nique such as SA or TS. GRASP can be
means of a heuristic criterion that gives effective if two conditions are satisfied:
them a score as a function of the (my-
opic) benefit if inserted in the current par- —the solution construction mechanism
tial solution. The candidate list, called re- samples the most promising regions of
stricted candidate list (RCL), is composed the search space;
of the best α elements. The heuristic val- —the solutions constructed by the con-
ues are updated at each step, thus the structive heuristic belong to basins of
scores of elements change during the con- attraction of different locally minimal
struction phase, depending on the possi- solutions;
ble choices. This constructive heuristic is
The first condition can be met by the
called dynamic, in contrast to the static
choice of an effective constructive heuris-
one which assigns a score to elements only
tic and an appropriate length of the can-
before starting the construction. For in-
didate list, whereas the second condition
stance, one of the static heuristics for the
can be met by choosing the constructive
TSP is based on arc costs: the lower the
heuristic and the local search in a way
cost of an arc, the higher its score. An ex-
such that they fit well.
ample of a dynamic heuristic is the cheap-
The description of GRASP as given
est insertion heuristic, where the score of
above indicates that a basic GRASP does
an element is evaluated depending on the
not use the history of the search process.11
current partial solution.
The only memory requirement is for stor-
The length α of the restricted candi-
ing the problem instance and for keep-
date list determines the strength of the
ing the best so-far solution. This is one
heuristic bias. In the extreme case of α = 1
of the reasons why GRASP is often out-
the best element would be added, thus
performed by other metaheuristics. How-
the construction would be equivalent to
ever, due to its simplicity, it is generally
a deterministic Greedy Heuristic. On the
very fast and it is able to produce quite
opposite, in case α = n the construction
good solutions in a very short amount of
would be completely random (indeed, the
computation time. Furthermore, it can be
choice of an element from the candidate
successfully integrated into other search
list is done at random). Therefore, α is
techniques. Among the applications of
a critical parameter which influences the
GRASP, we mention the JSS problem
sampling of the search space. In Pitsoulis
and Resende [2002] the most important
schemes to define α are listed. The sim- 11 However, some extensions in this direction are

plest scheme is, trivially, to keep α con- cited in Pitsoulis and Resende [2002], and an ex-
ample for a metaheuristic method using an adap-
stant; it can also be changed at each iter- tive greedy procedure depending on search history
ation, either randomly or by means of an is Squeaky Wheel Optimization (SWO) [Joslin and
adaptive scheme. Clements 1999].

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 279

Select a set of neighborhood structures Nk , k = 1, . . . , kmax becomes the local search starting point.
s ← GenerateInitialSolution()
while termination conditions not met do
The local search can use any neighbor-
k←1 hood structure and is not restricted to the
while k < kmax do % Inner loop set of neighborhood structures Nk , k =
s0 ← PickAtRandom(Nk (s))
00
% Shaking phase 1, . . . , kmax . At the end of the local search
s ← LocalSearch(s0 ) process (terminated as soon as a prede-
00
if ( f (s ) < f (s)) then
00 fined termination condition is verified) the
s←s
k←1
new solution s00 is compared with s and, if
else it is better, it replaces s and the algorithm
k ←k+1 starts again with k = 1. Otherwise, k is in-
endif cremented and a new shaking phase starts
endwhile
using a different neighborhood.
endwhile
The objective of the shaking phase is
Fig. 7. Algorithm: Variable Neighborhood Search to perturb the solution so as to provide a
(VNS). good starting point for the local search.
The starting point should belong to the
[Binato et al. 2001], the graph pla- basin of attraction of a different local min-
narization problem [Resende and Ribeiro imum than the current one, but should
1997] and assignment problems [Prais not be “too far” from s, otherwise the al-
and Ribeiro 2000]. A detailed and anno- gorithm would degenerate into a simple
tated bibliography references many more random multi-start. Moreover, choosing s0
applications [Festa and Resende 2002]. in the neighborhood of the current best so-
lution is likely to produce a solution that
maintains some good features of the cur-
3.4.2. Variable Neighborhood Search. Vari-
rent one.
able Neighborhood Search (VNS) is a
The process of changing neighborhoods
metaheuristic proposed in Hansen and
in case of no improvements corresponds
Mladenović [1999, 2001], which explicitly
to a diversification of the search. In par-
applies a strategy based on dynamically
ticular the choice of neighborhoods of
changing neighborhood structures. The al-
increasing cardinality yields a progres-
gorithm is very general and many degrees
sive diversification. The effectiveness of
of freedom exist for designing variants and
this dynamic neighborhood strategy can
particular instantiations.12
be explained by the fact that a “bad”
At the initialization step, a set of neigh-
place on the search landscape given by
borhood structures has to be defined.
one neighborhood could be a “good” place
These neighborhoods can be arbitrarily
on the search landscape given by an-
chosen, but often a sequence |N1 | < |N2 | <
other neighborhood.14 Moreover, a solu-
· · · < |Nkmax | of neighborhoods with increas-
tion that is locally optimal with respect
ing cardinality is defined.13 Then an initial to a neighborhood is probably not locally
solution is generated, the neighborhood optimal with respect to another neighbor-
index is initialized and the algorithm iter- hood. These concepts are known as “One
ates until a stopping condition is met (see Operator, One Landscape” and explained
Figure 7). VNS’ main cycle is composed in Jones [1995a, 1995b]. The core idea
of three phases: shaking, local search and is that the neighborhood structure deter-
move. In the shaking phase a solution s0 mines the topological properties of the
in the kth neighborhood of the current search landscape, that is, each neighbor-
solution s is randomly selected. Then, s0 hood defines one landscape. The proper-
ties of a landscape are in general different
12 The variants described in the following are also de-
from those of other landscapes, therefore
scribed in Hansen and Mladenović [1999, 2001].
13 In principle they could be one included in the other,
N1 ⊂ N2 ⊂ · · · ⊂ Nkmax . Nevertheless, such a se-
quence might produce an inefficient search, because 14 A
“good” place in the search space is an area from
a large number of solutions could be revisited. which a good local minimum can be reached.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


280 C. Blum and A. Roli

Fig. 8. Two search landscapes defined by two different neighborhoods. On the landscape that is
shown in the graphic on the left, the best improvement local search stops at sˆ1 , while it proceeds
till a better local minimum sˆ2 on the landscape that is shown in the graphic on the right.

a search strategy performs differently on ture Nk is defined. Local search only re-
them (see an example in Figure 8). gards changes on the variables belonging
This property is directly exploited by to the sub-problem it is applied to. The in-
a local search called Variable Neighbor- ner loop of VNDS is the following:
hood Descent (VND). In VND a best im- s0 ← PickAtRandom(Nk (s)) % s and s0 differ in k
provement local search (see Section 3.1) attributes
is applied, and, in case a local minimum s00 ← LocalSearch(s0 ,Attributes) % only moves
is found, the search proceeds with an- involving the k
other neighborhood structure. The VND attributes are
algorithm can be obtained by substituting allowed
the inner loop of the VNS algorithm (see if ( f (s00 ) < f (s)) then
Figure 7) with the following pseudo-code: s ← s00
k←1
s0 ← ChooseBestOf(Nk (s)) else
0
if ( f (s ) < f (s)) k ←k+1
then % i.e., if a better solution is found in Nk (s) endif
0
s←s
else % i.e., s is a local minimum The decision whether to perform a move
k ←k+1 can be varied as well. The acceptance cri-
endif terion based on improvements is strongly
As can be observed from the description steepest descent-oriented and it might not
as given above, the choice of the neigh- be suited to effectively explore the search
borhood structures is the critical point of space. For example, when local minima
VNS and VND. The neighborhoods cho- are clustered, VNS can quickly find the
sen should exploit different properties and best optimum in a cluster, but it has no
characteristics of the search space, that is, guidance to leave that cluster and find an-
the neighborhood structures should pro- other one. Skewed VNS (SVNS) extends
vide different abstractions of the search VNS by providing a more flexible accep-
space. A variant of VNS is obtained by se- tance criterion that takes also into account
lecting the neighborhoods in such a way the distance from the current solution.15
The new acceptance criterion is the fol-
as to produce a problem decomposition
lowing: besides always accepting improve-
(the algorithm is called Variable Neigh-
ments, worse solutions can be accepted if
borhood Decomposition Search—VNDS).
the distance from the current one is less
VNDS follows the usual VNS scheme, but
than a value αρ(s, s00 ). The function ρ(s, s00 )
the neighborhood structures and the lo-
measures the distance between s and s00
cal search are defined on sub-problems.
For each solution, all attributes (usually
variables) are kept fixed except for k of 15 A distance measure between solutions has thus to
them. For each, k, a neighborhood struc- be formally defined.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 281

Fig. 9. Basic GLS idea: Escaping from a valley in the landscape by increas-
ing the objective function value of its solutions.

and α is a parameter that weights the The basic GLS principle is to help the
importance of the distance between the search to gradually move away from lo-
two solutions in the acceptance criterion. cal minima by changing the search land-
The inner loop of SVNS can be sketched as scape. In GLS, the set of solutions and
follows: the neighborhood structure are kept fixed,
00 00 while the objective function f is dynami-
if ( f (s ) − αρ(s, s ) < f (s)) then
s←s
00 cally changed with the aim of making the
k←1 current local optimum “less desirable”. A
else pictorial description of this idea is given in
k ←k+1 Figure 9.
endif The mechanism used by GLS is based
on solution features, which may be any
VNS and its variants have been success- kind of properties or characteristics that
fully applied to graph based CO problems can be used to discriminate between solu-
such as the p-Median problem [Hansen tions. For example, solution features in the
and Mladenović 1997], the degree con- TSP could be arcs between pairs of cities,
strained minimum spanning tree prob- while in the MAXSAT problem they could
lem [Ribeiro and Souza 2002], the Steiner be the number of unsatisfied clauses. An
tree problem [Wade and Rayward-Smith indicator function Ii (s) indicates whether
1997] and the k-Cardinality Tree (KCT) the feature i is present in solution s:
problem [Mladenović and Urošević 2001].
References to more applications can be (
1 : if feature i is present in
found in Hansen and Mladenović [2001].
Ii (s) = : solution s
0 : otherwise.
3.4.3. Guided Local Search. Tabu Search
and Variable Neighborhood Search ex- The objective function f is modified to
plicitly deal with dynamic neighborhoods 0
yield a new objective function f by adding
with the aim of efficiently and effectively a term that depends on the m features:
exploring the search space. A different
approach for guiding the search is to 0
X
m
dynamically change the objective func- f (s) = f (s) + λ pi · Ii (s),
tion. Among the most general methods i=1
that use this approach is Guided Local
Search (GLS) [Voudouris and Tsang 1999; where pi are called penalty parame-
Voudouris 1997]. ters and λ is called the regularization

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


282 C. Blum and A. Roli

s ← GenerateInitialSolution() date rules are often very sensitive to the


while termination conditions not met do problem instance.
0
s ← LocalSearch(s, f ) GLS has been successfully applied to
for all feature i with maximum utility Util(s, i) do
pi ← pi + 1
the weighted MAXSAT [Mills and Tsang
endfor 2000], the VR problem [Kilby et al. 1999],
Update( f 0 , p) % p is the penalty vector the TSP and the QAP [Voudouris and
endwhile Tsang 1999].
Fig. 10. Algorithm: Guided Local Search (GLS).
3.4.4. Iterated Local Search. We conclude
this presentation of explorative strate-
parameter. The penalty parameters gies with Iterated Local Search (ILS), the
weight the importance of the features: most general scheme among the explo-
the higher pi , the higher the importance rative strategies. On the one hand, its gen-
of feature i, thus the higher the cost of erality makes it a framework for other
having that feature in the solution. The metaheuristics (such as VNS); on the other
regularization parameter balances the re- hand, other metaheuristics can be eas-
levance of features with respect to the ily incorporated as subcomponents. ILS is
original objective function. a simple but powerful metaheuristic al-
The algorithm (see Figure 10) works as gorithm [Stützle 1999a,1999b; Lourenço
follows: It starts from an initial solution et al. 2001,2002; Martin et al. 1991]. It
and applies a local search method until a applies local search to an initial solution
local minimum is reached. Then the array until it finds a local optimum; then it per-
p = ( p1 , . . . , pm ) of penalties is updated turbs the solution and it restarts local
by incrementing some of the penalties and search. The importance of the perturba-
the local search is started again. The pe- tion is obvious: too small a perturbation
nalized features are those that have the might not enable the system to escape
maximum utility: from the basin of attraction of the local op-
timum just found. On the other side, too
ci strong a perturbation would make the al-
Util(s, i) = Ii (s) · , gorithm similar to a random restart local
1 + pi
search.
A local search is effective if it is able to
where ci are costs assigned to every feature find good local minima, that is, if it can
i giving a heuristic evaluation of the rel- find the basin of attraction of those states.
ative importance of features with respect When the search space is wide and/or
to others. The higher the cost, the higher when the basins of attraction of good local
the utility of features. Nevertheless, the optima are small,16 a simple multi-start
cost is scaled by the penalty parameter to algorithm is almost useless. An effective
prevent the algorithm from being totally search could be designed as a trajectory
biased toward the cost and to make it sen-
only in the set of local optima Ŝ, instead
sitive to the search history.
of in the set S of all the states. Unfortu-
The penalties update procedure can be
nately, in most cases there is no feasible
modified by adding a multiplicative rule
way of introducing a neighborhood struc-
to the simple incrementing rule (that is
applied at each iteration). The multiplica- ture for Ŝ. Therefore, a trajectory along
tive rule has the form: pi ← pi · α, where local optima sˆ1 , sˆ2 , . . . , sˆt is performed,
α ∈ (0, 1). This rule is applied with a lower without explicitly introducing a neighbor-
frequency than the incrementing one (for hood structure, by applying the following
example every few hundreds of iterations) scheme:
with the aim of smoothing the weights of
penalized features so as to prevent the 16 The basin of attraction size of a point s (in a finite
landscape from becoming too rugged. It is space), is defined as the fraction of initial states of
important to note that the penalties up- trajectories which converge to point s.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 283

s0 ← GenerateInitialSolution() that an initial solution is considered a good


ŝ ← LocalSearch(s0 ) starting point depending on the particu-
while termination conditions not met do lar LS applied and on the problem struc-
s0 ← Perturbation(ŝ, history)
ture, thus the algorithm designer’s goal is
sˆ0 ← LocalSearch(s0 )
to find a trade-off between speed and qual-
ŝ ← ApplyAcceptanceCriterion(ŝ, sˆ0 , history)
endwhile
ity of solutions.
The perturbation is usually non-
Fig. 11. Algorithm: Iterated Local Search deterministic in order to avoid cycling.
(ILS). Its most important characteristic is the
strength, roughly defined as the amount of
(1) Execute local search (LS) from an ini-
changes made on the current solution. The
tial state s until a local minimum ŝ is
strength can be either fixed or variable.
found.
In the first case, the distance between
(2) Perturb ŝ and obtain s0 . ŝ and s0 is kept constant, independently
(3) Execute LS from s0 until a local mini- of the problem size. However, a variable
mum ŝ0 is reached. strength is in general more effective,
(4) On the basis of an acceptance criterion since it has been experimentally found
decide whether to set ŝ ← ŝ0 . that, in most of the problems, the bigger
(5) Goto step 2. the problem size, the larger should be
the strength. More sophisticated schemes
The requirement on the perturbation of ŝ is are possible; for example, the strength
to produce a starting point for local search can be adaptive: it increases when more
such that a local minimum different from diversification is needed and it decreases
ŝ is reached. However, this new local min- when intensification seems preferable.
imum should be closer to ŝ than a local VNS and its variants belong to this cat-
minimum produced by a random restart. egory. A second choice is the mechanism
The acceptance criterion acts as a counter- to perform perturbations. This may be a
balance, as it filters and gives feedback to random mechanism, or the perturbation
the perturbation action, depending on the may be produced by a (semi-)deterministic
characteristics of the new local minimum. method (e.g., a LS different from the one
A high level description of ILS as it is de- used in the main algorithm).
scribed in Lourenço et al. [2002] is given The third important component is the
in Figure 11. Figure 12 shows a possible acceptance criterion. Two extreme exam-
(lucky) ILS step. ples consist in (1) accepting the new local
The design of ILS algorithms has sev- optimum only in case of improvement and
eral degrees of freedom in the choice of (2) in always accepting the new solution.
the initial solution, perturbation and ac- In-between, there are several possibilities.
ceptance criteria. A key role is played by For example, it is possible to adopt a kind
the history of the search which can be ex- of annealing schedule: accept all the im-
ploited both in form of short and long term proving new local optima and accept also
memory. the nonimproving ones with a probability
The construction of initial solutions that is a function of the temperature T and
should be fast (computationally not ex- the difference of objective function values.
pensive), and initial solutions should be a In formulas:
good starting point for local search. The
fastest way of producing an initial solu- p(Accept(ŝ, ŝ0 , history))
tion is to generate it at random; how- 
 1 if f (ŝ0 )< f (ŝ)
ever, this is the easiest way for problems µ 0 ¶
that are constrained, whilst in other cases = f (ŝ ) − f (ŝ)
exp − otherwise
the construction of a feasible solution re- T
quires also constraint checking. Construc-
tive methods, guided by heuristics, can The cooling schedule can be either
also be adopted. It is worth underlining monotonic (non-increasing in time) or

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


284 C. Blum and A. Roli

Fig. 12. A desirable ILS step: the local minimum ŝ is perturbed, then LS
is applied and a new local minimum is found.

non-monotonic (adapted to tune the bal- based methods in combinatorial optimiza-


ance between diversification and intensifi- tion are Evolutionary Computation (EC)
cation). The nonmonotonic schedule is par- and Ant Colony Optimization (ACO). In
ticularly effective if it exploits the history EC algorithms, a population of individu-
of the search, in a way similar to the Re- als is modified by recombination and mu-
active Tabu Search [Taillard 1991] men- tation operators, and in ACO a colony of
tioned at the end of the section about Tabu artificial ants is used to construct solu-
Search. When intensification seems no tions guided by the pheromone trails and
longer effective, a diversification phase is heuristic information.
needed and the temperature is increased.
Examples for successful applications of
ILS are to the TSP [Martin and Otto 4.1. Evolutionary Computation
1996; Johnson and McGeoch 1997], to Evolutionary Computation (EC) algo-
the QAP [Lourenço et al. 2002], and to rithms are inspired by nature’s capabil-
the Single Machine Total Weighted Tardi- ity to evolve living beings well adapted
ness (SMTWT) problem [den Besten et al. to their environment. EC algorithms can
2001]. References to other applications be succinctly characterized as computa-
can be found in Lourenço et al. [2002]. tional models of evolutionary processes. At
each iteration a number of operators is ap-
4. POPULATION-BASED METHODS plied to the individuals of the current po-
Population-based methods deal in every it- pulation to generate the individuals of the
eration of the algorithm with a set (i.e., population of the next generation (itera-
a population) of solutions17 rather than tion). Usually, EC algorithms use opera-
with a single solution. As they deal with a tors called recombination or crossover to
population of solutions, population-based recombine two or more individuals to pro-
algorithms provide a natural, intrinsic duce new individuals. They also use muta-
way for the exploration of the search tion or modification operators which cause
space. Yet, the final performance depends a self-adaptation of individuals. The driv-
strongly on the way the population is ma- ing force in evolutionary algorithms is the
nipulated. The most studied population- selection of individuals based on their fit-
ness (this can be the value of an objec-
17 In general, especially in EC algorithms, we talk tive function or the result of a simulation
about a population of individuals rather than experiment, or some other kind of qual-
solutions. ity measure). Individuals with a higher

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 285

fitness have a higher probability to be cho- P ← GenerateInitialPopulation()


sen as members of the population of the Evaluate(P )
next iteration (or as parents for the gen- while termination conditions not met do
P 0 ← Recombine(P )
eration of new individuals). This corre-
P 00 ← Mutate(P 0 )
sponds to the principle of survival of the Evaluate(P 00 )
fittest in natural evolution. It is the capa- P ← Select(P 00 ∪ P )
bility of nature to adapt itself to a chang- endwhile
ing environment, which gave the inspira-
tion for EC algorithms. Fig. 13. Algorithm: Evolutionary Com-
putation (EC).
There has been a variety of slightly dif-
ferent EC algorithms proposed over the
years. Basically they fall into three differ- offspring population. The main features
ent categories which have been developed of an EC algorithm are outlined in the
independently from each other. These are following.
Evolutionary Programming (EP) devel-
oped by Fogel [1962] and Fogel et al. Description of the Individuals. EC algo-
[1966], Evolutionary Strategies (ES) pro- rithms handle populations of individuals.
posed by Rechenberg [1973] and Genetic These individuals are not necessarily solu-
Algorithms initiated by Holland [1975] tions of the considered problem. They may
(see Goldberg [1989], Mitchell [1998], be partial solutions, or sets of solutions,
Reeves and Rowe [2002], and Vose [1999] or any object which can be transformed
for further references). EP arose from the into one or more solutions in a structured
desire to generate machine intelligence. way. Most commonly used in combinato-
While EP originally was proposed to op- rial optimization is the representation of
erate on discrete representations of finite solutions as bit-strings or as permutations
state machines, most of the present vari- of n integer numbers. Tree-structures or
ants are used for continuous optimization other complex structures are also possible.
problems. The latter also holds for most In the context of Genetic Algorithms, indi-
present variants of ES, whereas GAs are viduals are called genotypes, whereas the
mainly applied to solve combinatorial op- solutions that are encoded by individuals
timization problems. Over the years, there are called phenotypes. This is to differen-
have been quite a few overviews and sur- tiate between the representation of solu-
veys about EC methods. Among those are tions and solutions themselves. The choice
the ones by Bäck [1996], Fogel [1994], of an appropriate representation is cru-
Spears et al. [1993] and Michalewicz and cial for the success of an EC algorithm.
Michalewicz [1997]. Calégary et al. [1999] Holland’s [1975] schema analysis and
propose a taxonomy of EC algorithms. Radcliffe’s [1991] generalization to formae
In the following we provide a “combina- are examples of how theory can help to
torial optimization”-oriented introduction guide representation choices.
to EC algorithms. For doing this, we fol- Evolution Process. In each iteration, it
low an overview work by Hertz and Kobler has to be decided which individuals will
[2000], which gives, in our opinion, a good enter the population of the next iteration.
overview of the different components of This is done by a selection scheme. To
EC algorithms and of the possibilities to choose, the individuals for the next po-
define them. pulation exclusively from the offspring is
Figure 13 shows the basic structure of called generational replacement. If it is
every EC algorithm. In this algorithm, possible to transfer individuals of the cur-
P denotes the population of individuals. rent population into the next population,
A population of offspring is generated then we deal with a so-called steady state
by the application of recombination and evolution process.
mutation operators and the individuals Most EC algorithms work with pop-
for the next population are selected from ulations of fixed size keeping at least
the union of the old population and the the best individual always in the current

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


286 C. Blum and A. Roli

population. It is also possible to have a an individual is sometimes more appropri-


variable population size. In case of a con- ate (or even unavoidable). The third pos-
tinuously shrinking population size, the sibility consists in trying to repair an in-
situation where only one individual is left feasible solution (see Eiben and Ruttkay
in the population (or no crossover partners [1997] for an example).
can be found for any member of the popu- Intensification Strategy. In many appli-
lation) might be one of the stopping condi- cations it proved to be quite beneficial to
tions of the algorithm. use improvement mechanisms to improve
Neighborhood Structure. A neighbor- the fitness of individuals. EC algorithms
hood function N EC : I → 2I on the set of that apply a local search algorithm to ev-
individuals I assigns to every individual ery individual of a population are often
i ∈ I a set of individuals N EC (i) ⊆ I which called Memetic Algorithms [Moscato 1989,
are permitted to act as recombination 1999]. While the use of a population en-
partners for i to create offspring. If an indi- sures an exploration of the search space,
vidual can be recombined with any other the use of local search techniques helps to
individual (as for example in the simple quickly identify “good” areas in the search
GA) we talk about unstructured popula- space.
tions, otherwise we talk about structured Another intensification strategy is
populations. An example for an EC al- the use of recombination operators that
gorithm that works on structured popu- explicitly try to combine “good” parts of in-
lations is the Parallel Genetic Algorithm dividuals (rather than, e.g., a simple one-
proposed by Mühlenbein [1991]. point crossover for bit-strings). This may
Information Sources. The most com- guide the search performed by EC algo-
mon form of information sources to cre- rithms to areas of individuals with certain
ate offspring (i.e., new individuals) is a “good” properties. Techniques of this kind
couple of parents (two-parent crossover). are sometimes called linkage learning
But there are also recombination oper- or building block learning (see Goldberg
ators that operate on more than two et al. [1991], van Kemenade [1996],
individuals to create a new individual Watson et al. [1998], and Harik [1999] as
(multi-parent crossover), see Eiben et al. examples). Moreover, generalized recom-
[1994]. More recent developments even bination operators have been proposed in
use population statistics for generat- the literature, which incorporate the no-
ing the individuals of the next popu- tion of “neighborhood search” into EC. An
lation. Examples are the recombination example can be found in Rayward-Smith
operators called Gene Pool Recombina- [1994].
tion [Mühlenbein and Voigt 1995] and Diversification Strategy. One of the ma-
Bit-Simulated Crossover [Syswerda 1993] jor difficulties of EC algorithms (especially
which make use of a distribution over the when applying local search) is the prema-
search space given by the current popula- ture convergence toward sub-optimal solu-
tion to generate the next population. tions. The most simple mechanism to di-
Infeasibility. An important characteris- versify the search process is the use of a
tic of an EC algorithm is the way it deals mutation operator. The simple form of a
with infeasible individuals. When recom- mutation operator just performs a small
bining individuals, the offspring might be random perturbation of an individual, in-
potentially infeasible. There are basically troducing a kind of noise. In order to avoid
three different ways to handle such a sit- premature convergence there are ways
uation. The most simple action is to re- of maintaining the population diversity.
ject infeasible individuals. Nevertheless, Probably the oldest strategies are crowd-
for many problems (e.g., for timetabling ing [DeJong 1975] and its close relative,
problems) it might be very difficult to find preselection. Newer strategies are fitness
feasible individuals. Therefore, the strat- sharing [Goldberg and Richardson 1987],
egy of penalizing infeasible individuals in respectively niching, whereby the repro-
the function that measures the quality of ductive fitness allocated to an individual

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 287

Initial Phase: Scatter Search (respectively, Path Re-


SeedGeneration() linking) is a search strategy that gener-
repeat ates a set of solutions from a chosen set
DiversificationGenerator()
of reference solutions corresponding to fea-
Improvement()
ReferenceSetUpdate()
sible solutions to the problem under con-
until the reference set is of cardinality n sideration. This is done by making com-
binations of subsets of the current set of
Scatter Search/Path Relinking Phase: reference solutions. The resulting solu-
repeat tions are called trial solutions. These trial
SubsetGeneration() solutions may be infeasible solutions and
SolutionCombination() are therefore usually modified by means of
Improvement() a repair procedure that transforms them
ReferenceSetUpdate() into feasible solutions. An improvement
until termination criteria met
mechanism is then applied in order to try
Fig. 14. Algorithm: Scatter Search and to improve the set of trial solutions (usu-
Path Relinking. ally this improvement procedure is a lo-
cal search). These improved solutions form
in a population is reduced proportion- the set of dispersed solutions. The new set
ally to the number of other individuals of reference solutions that will be used
that share the same region of the search in the next iteration is selected from the
space. current set of reference solutions and the
This concludes the list of the main newly created set of dispersed solutions.
features of EC algorithms. EC algorithms The components of the pseudo-code, which
have been applied to most CO problems is shown in Figure 14, are explained in the
and optimization problems in general. following:
Recent successes were obtained in the SeedGeneration(): One or more seed so-
rapidly growing bioinformatics area (see, lutions, which are arbitrary trial solu-
e.g., Fogel et al. [2002]), but also in multi- tions, are created and used to initiate the
objective optimization [Coello Coello remainder of the method.
2000], and in evolvable hardware [Sipper DiversificationGenerator(): This is a pro-
et al. 1997]. For an extensive collection cedure to generate a collection of diverse
of references to EC applications, we refer trial solutions from an arbitrary trial so-
to Bäck et al. [1997]. In the following two lution (or seed solution).
sections, we are going to introduce two Improvement(): In this procedure, an im-
other populations-based methods that provement mechanism—usually a local
are sometimes also regarded as being EC search—is used to transform a trial solu-
algorithms. tion into one or more enhanced trial solu-
tions. Neither the input nor the output so-
4.1.1. Scatter Search and Path Relinking. lutions are required to be feasible, though
Scatter Search and its generalized form the output solutions will more usually be
called Path Relinking [Glover 1999; expected to be so. It might be necessary
Glover et al. 2000] differ from EC al- to apply repair methods to infeasible solu-
gorithms mainly by providing unifying tions.
principles for joining (or recombining) so- ReferenceSetUpdate(): The procedure
lutions based on generalized path con- for updating the reference set is respon-
structions in Euclidean or neighborhood sible for building and maintaining a refer-
spaces. They also incorporate some ideas ence set consisting of a number of “best”
originating from Tabu Search methods, as, solutions found in the course of the al-
for example, the use of adaptive memory gorithm. The attribute “best” covers fea-
and associated memory-exploiting mech- tures such as quality of solutions and di-
anisms. The template for Scatter Search versity of solutions (the solutions in the
(respectively, Path Relinking) is shown in reference set should be of good quality and
Figure 14. they should be diverse).

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


288 C. Blum and A. Roli

SubsetGeneration(): This method oper- P ← InitializePopulation()


ates on the reference set, to produce a sub- while termination criteria not met do
Psel ← Select( P ) % Psel ⊆ P
set of its solutions as a basis for creating
p(x) = p(x | Psel ) ← EstimateProbabilityDistribution()
combined solutions. P ← SampleProbabilityDistribution()
SolutionCombination(): A procedure to endwhile
transform a given subset of solutions pro-
duced by the subset generation method Fig. 15. Algorithm: Estimation of Distribution Al-
gorithms (EDAs).
into one or more combined solutions. In
Scatter Search, which was introduced for break good building blocks. So, a num-
solutions encoded as points in the Eu- ber of algorithms—sometimes called Esti-
clidean space, new solutions are created mation of Distribution Algorithms (EDA)
by building linear combinations of refer- [Mühlenbein and Paaß 1996]—have been
ence solutions using both positive and neg- developed (see Figure 15 for the algorith-
ative weights. This means that trial solu- mic framework). These algorithms, which
tions can be both, inside and outside the have a theoretical foundation in probabil-
convex region spanned by the reference ity theory, are also based on populations
solutions. In Path Relinking, the concept that evolve as the search progresses. EDAs
of combining solutions by making linear use probabilistic modelling of promising
combinations of reference points is gen- solutions to estimate a distribution over
eralized to neighborhood spaces. Linear the search space, which is then used to pro-
combinations of points in the Euclidean duce the next generation by sampling the
space can be re-interpreted as paths be- search space according to the estimated
tween and beyond solutions in a neigh- distribution. After every iteration, the dis-
borhood space. To generate the desired tribution is re-estimated. For a survey of
paths, it is only necessary to select moves EDAs, see [Pelikan et al. 1999b].
that satisfy the following condition: upon One of the first EDAs that was proposed
starting from an initiating solution, the for Combinatorial Optimization is called
moves must progressively introduce at- Population-based Incremental Learning
tributes contributed by a guiding solution. (PBIL) [Baluja 1994; Baluja and Caruana
Multiparent path generation possibilities 1995]. The objective of this method is to
emerge in Path Relinking by consider- create a real valued probability vector
ing the combined attributes provided by (each position corresponds to a binary
a set of guiding solutions, where these at- decision variable) which—when used to
tributes are weighted to determine which sample the search space—generates high
moves are given higher priority. quality solutions with high probability.
Scatter Search enjoys increasing inter- Initially, the values of the probability vec-
est in recent years. Among other prob- tor are initialized to 0.5 (for each variable
lems, it has been applied to multi- there is equal probability to be set to 0 or
objective assignment problems [Laguna 1). The goal of shifting the values of this
et al. 2000] and the Linear Ordering Prob- probability vector in order to generate
lem (LOP) [Campos et al. 2001]. For fur- high quality solutions is accomplished as
ther references, we refer to [Glover et al. follows: a number of solution vectors are
2002]. Path relinking is often used as generated according to the probability vec-
a component in metaheuristics such as tor. Then, the probability vector is shifted
Tabu Search [Laguna et al. 1999] and toward the generated solution vector(s)
GRASP [Aiex et al. 2003; Laguna and with highest quality. The distance that
Martı́ 1999]. the probability vector is shifted depends
on the learning rate parameter. Then,
4.1.2. Estimation of Distribution Algorithms. a mutation operator is applied to the
In the last decade, more and more re- probability vector. After that, the cycle is
searchers tried to overcome the draw- repeated. The probability vector can
backs of usual recombination operators be regarded as a prototype vector for
of EC algorithms, which are likely to generating high-quality solution vectors

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 289

with respect to the available knowledge components to a partial solution under


about the search space. The drawback consideration.18 For doing that, artificial
of this method is the fact that it does ants perform randomized walks on a com-
not automatically provide a way to deal pletely connected graph G = (C, L) whose
with constrained problems. In contrast vertices are the solution components C
to PBIL, which estimates a distribution and the set L are the connections. This
of promising solutions assuming that graph is commonly called construction
the decision variables are independent, graph. When a constrained CO problem
various other approaches try to estimate is considered, the problem constraints
distributions taking into account depen- Ä are built into the ants’ constructive
dencies between decision variables. An procedure in such a way that in every step
example for EDAs regarding pairwise of the construction process only feasible
dependencies between decision variables solution components can be added to
is MIMIC [de Bonet et al. 1997] and an the current partial solution. In most
example for multivariate dependencies applications, ants are implemented to
is the Bayesian Optimization Algorithm build feasible solutions, but sometimes it
(BOA) [Pelikan et al. 1999a]. is unavoidable to also let them construct
The field of EDAs is still quite young infeasible solutions. Components ci ∈ C
and much of the research effort is fo- can have associated a pheromone trail
cused on methodology rather than high- parameter Ti , and connections l i j ∈ L can
performance applications. Applications to have associated a pheromone trail param-
Knapsack problems, the Job Shop Schedu- eter Ti j . The set of all pheromone trail
ling (JSS) problem, and other CO prob- parameters is denoted by T . The values
lems can be found in Larrañaga and of these parameters—the pheromone
Lozano [2002]. values—are denoted by τi , respectively
τi j . Furthermore, components and connec-
tions can have associated a heuristic value
4.2. Ant Colony Optimization
ηi , respectively ηij , representing a priori
Ant Colony Optimization (ACO) is a meta- or run time heuristic information about
heuristic approach proposed in Dorigo the problem instance. We henceforth
1992, 1996, 1999. In the course of this denote the set of all heuristic values by H.
section, we keep close to the description These values are used by the ants to make
of ACO as given in Dorigo and Di Caro probabilistic decisions on how to move on
[1999]. The inspiring source of ACO is the construction graph. The probabilities
the foraging behavior of real ants. This involved in moving on the construction
behavior—as described by Deneubourg graph are commonly called transition
et al. [1990]—enables ants to find short- probabilities. The first ACO algorithm
est paths between food sources and their proposed in the literature is called Ant
nest. While walking from food sources to System (AS) [Dorigo et al. 1996]. The
the nest and vice versa, ants deposit a sub- pseudo-code for this algorithm is shown
stance called pheromone on the ground. in Figure 16. For the sake of simplicity, we
When they decide about a direction to go, restrict the following description of AS to
they choose with higher probability paths pheromone trail parameters and heuristic
that are marked by stronger pheromone information on solution components.
concentrations. This basic behavior is the In this algorithm, A denotes the set
basis for a cooperative interaction which of ants and sa denotes the solution con-
leads to the emergence of shortest paths. structed by ant a ∈ A. After the initial-
ACO algorithms are based on a ization of the pheromone values, at each
parametrized probabilistic model—the step of the algorithm each ant constructs a
pheromone model—that is used to model
the chemical pheromone trails. Artificial 18 Therefore, the ACO metaheuristic can be applied to
ants incrementally construct solutions any CO problem for which a constructive procedure
by adding opportunely defined solution can be defined.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


290 C. Blum and A. Roli

InitializePheromoneValues(T ) while termination conditions not met do


while termination conditions not met do ScheduleActivities
for all ants a ∈ A do AntBasedSolutionConstruction()
sa ← ConstructSolution(T ,H) PheromoneUpdate()
endfor DaemonActions() % optional
ApplyOnlineDelayedPheromoneUpdate(T ,{sa | a ∈ A}) end ScheduleActivities
endwhile
endwhile
Fig. 16. Algorithm: Ant System (AS).
Fig. 17. Algorithm: Ant Colony Opti-
solution. These solutions are then used to mization (ACO).
update the pheromone values. The compo-
nents of this algorithm are explained in where F : S 7→ IR + is a function that satis-
more detail in the following. fies f (s) < f (s0 ) ⇒ F (s) ≥ F (s0 ), ∀s 6= s0 ∈
InitializePheromoneValues(T ): At the be- S. F (·) is commonly called the quality
ginning of the algorithm the pheromone function. Furthermore, 0 < ρ ≤ 1 is the
values are initialized to the same small pheromone evaporation rate. This phero-
value ph > 0. mone update rule aims at an increase of
ConstructSolution(T ,H): In the construc- pheromone on solution components that
tion phase an ant incrementally builds a have been found in high-quality solutions.
solution by adding solution components to In the following, we describe the more
the partial solution constructed so far. The general ACO metaheuristic, which is
probabilistic choice of the next solution based on the same basic principles as AS.
component to be added is done by means The ACO metaheuristic framework that is
of transition probabilities, which in AS are shown in Figure 17 covers all the improve-
determined by the following state transi- ments and extensions of AS which have
tion rule: been developed over the years. It consists
p(cr |sa [cl ]) of three parts gathered in the Schedule-
 Activities construct. The ScheduleActivities

P [ηr ]α [τr ]β construct does not specify how these three
α β
if cr ∈ J (sa [cl ])
= cu ∈J (sa [cl ]) [ηu ] [τu ]
activities are scheduled and synchronized.

 This is up to the algorithm designer.
0 otherwise
AntBasedSolutionConstruction(): An ant
(1) constructively builds a solution to the
problem by moving through nodes of the
In this formula, α and β are parameters to construction graph G. Ants move by apply-
adjust the relative importance of heuris- ing a stochastic local decision policy that
tic information and pheromone values and makes use of the pheromone values and
J (sa [cl ]) denotes the set of solution compo- the heuristic values on components and/or
nents that are allowed to be added to the connections of the construction graph (e.g.,
partial solution sa [cl ], where cl is the last see the state transition rule of AS). While
component that was added. moving, the ant keeps in memory the par-
ApplyOnlineDelayedPheromoneUpdate tial solution it has built in terms of the
(T ,{sa | a ∈ A}): Once all ants have con- path it was walking on the construction
structed a solution, the online delayed graph.
pheromone update rule is applied: PheromoneUpdate(): When adding a
X component c j to the current partial so-
τ j ← (1 − ρ) · τ j + 1τ sj a (2) lution, an ant can update the pheromone
a∈A trail(s) τi and/or τi j (in case the ant was
walking on connection l i j in order to reach
∀ T j ∈ T , where component c j ). This kind of pheromone
½ update is called online step-by-step phero-
sa F (sa ) if c j is a component of sa
1τ j = mone update. Once an ant has built a so-
0 otherwise, lution, it can retrace the same path back-
(3) ward (by using its memory) and update

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 291

the pheromone trails of the used compo- rule. With this rule, some moves are cho-
nents and/or connections according to the sen deterministically (in a greedy man-
quality of the solution it has built. This is ner), others are chosen probabilistically
called online delayed pheromone update. with the usual decision rule. Third, in
Pheromone evaporation is the process by ACS, ants perform only online step-
means of which the pheromone trail in- by-step pheromone updates. These up-
tensity on the components decreases over dates are performed to favor the emer-
time. From a practical point of view, phe- gence of other solutions than the best so
romone evaporation is needed to avoid a far.
too rapid convergence of the algorithm to- MAX -MIN Ant System (MMAS).
ward a sub-optimal region. It implements MMAS is also an extension of AS. First,
a useful form of forgetting, favoring the ex- the pheromone trails are only updated of-
ploration of new areas in the search space. fline by the daemon (the arcs that were
DaemonActions(): Daemon actions can used by the iteration best ant or the best
be used to implement centralized actions ant since the start of the algorithm receive
which cannot be performed by single ants. additional pheromone). Second, the phero-
Examples are the use of a local search pro- mone values are restricted to an interval
cedure applied to the solutions built by the [τmin , τmax ] and the pheromone trails are
ants, or the collection of global informa- initialized to their maximum value τmax .
tion that can be used to decide whether it Explicit bounds on the pheromone trails
is useful or not to deposit additional phe- prevent that the probability to construct a
romone to bias the search process from solution falls below a certain value greater
a nonlocal perspective. As a practical ex- than 0. This means that the chance of find-
ample, the daemon can observe the path ing a global optimum never vanishes dur-
found by each ant in the colony and choose ing the course of the algorithm.
to deposit extra pheromone on the compo- Recently, researchers have been deal-
nents used by the ant that built the best ing with finding similarities between
solution. Pheromone updates performed ACO algorithms and probabilistic learn-
by the daemon are called offline phero- ing algorithms such as EDAs. An im-
mone updates. portant step into this direction was the
Within the ACO metaheuristic frame- development of the Hyper-Cube Frame-
work, as shortly described above, the cur- work for Ant Colony Optimization (HC-
rently best performing versions in prac- ACO) [Blum et al. 2001]. An extensive
tise are Ant Colony System (ACS) [Dorigo study on this subject has been presented
and Gambardella 1997] and MAX -MIN in Zlochin et al. [2004], where the au-
Ant System (MMAS) [Stützle and Hoos thors present a unifying framework for
2000]. In the following, we are going to so-called Model-Based Search (MBS) al-
briefly outline the peculiarities of these al- gorithms. Also, the close relation of algo-
gorithms. rithms like Population-Based Incremen-
Ant Colony System (ACS). The ACS al- tal Learning (PBIL) [Baluja and Caruana
gorithm has been introduced to improve 1995] and the Univariate Marginal Dis-
the performance of AS. ACS is based on tribution Algorithm (UMDA) [Mühlenbein
AS but presents some important differ- and Paaß 1996] to ACO algorithms in the
ences. First, the daemon updates phero- Hyper-Cube Framework has been shown.
mone trails offline: At the end of an iter- We refer the interested reader to Zlochin
ation of the algorithm—once all the ants et al. [2004] for more information on this
have built a solution—pheromone is added subject. Furthermore, connections of ACO
to the arcs used by the ant that found the algorithms to Stochastic Gradient Descent
best solution from the start of the algo- (SGD) algorithms are shown in Meuleau
rithm. Second, ants use a different deci- and Dorigo [2002].
sion rule to decide to which component to Successful applications of ACO include
move next in the construction graph. The the application to routing in communi-
rule is called pseudo-random-proportional cation networks [Di Caro and Dorigo

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


292 C. Blum and A. Roli

1998], the application to the Sequential fication). However, the terms exploitation
Ordering Problem (SOP) [Gambardella and exploration have a somewhat more re-
and Dorigo 2000], and the application to stricted meaning. In fact, the notions of
Resource Constraint Project Scheduling exploitation and exploration often refer to
(RCPS) [Merkle et al. 2002]. Further refer- rather short term strategies tied to ran-
ences to applications of ACO can be found domness, whereas intensification and di-
in Dorigo and Stützle [2002, 2003]. versification refer to rather medium and
long term strategies based on the usage of
memory. As the various different ways of
5. A UNIFYING VIEW ON INTENSIFICATION
using memory become increasingly impor-
AND DIVERSIFICATION
tant in the whole field of metaheuristics,
In this section, we take a closer look at the terms intensification and diversifica-
the concepts of intensification and diversi- tion are more and more adopted and un-
fication as the two powerful forces driving derstood in their original meaning.
metaheuristic applications to high perfor- An implicit reference to the concept of
mance. We give a view on metaheuristics “locality” is often introduced when inten-
that is characterized by the way inten- sification and diversification are involved.
sification and diversification are imple- The notion of “area” (or “region”) of the
mented. Although the relevance of these search space and of “locality” can only be
two concepts is commonly agreed, so far expressed in a fuzzy way, as they always
there is no unifying description to be found depend on the characteristics of the search
in the literature. Descriptions are very space as well as on the definition of metrics
generic and metaheuristic specific. There- on the search space (distances between
fore most of them can be considered incom- solutions).
plete and sometimes they are even oppos- The literature provides several high
ing. Depending on the paradigm behind level descriptions of intensification and
a particular metaheuristic, intensification diversification. In the following, we cite
and diversification are achieved in differ- some of them.
ent ways. Even so, we propose a unifying
view on intensification and diversification. “Two highly important components
Furthermore, this discussion could lead to of Tabu Search are intensification and
the goal-directed development of hybrid diversification strategies. Intensification
algorithms combining concepts originat- strategies are based on modifying choice
ing from different metaheuristics. rules to encourage move combinations
and solution features historically found
good. They may also initiate a return to
5.1. Intensification and Diversification
attractive regions to search them more
Every metaheuristic approach should be thoroughly. Since elite solutions must be
designed with the aim of effectively and recorded in order to examine their imme-
efficiently exploring a search space. The diate neighborhoods, explicit memory is
search performed by a metaheuristic ap- closely related to the implementation of
proach should be “clever” enough to both intensification strategies. The main dif-
intensively explore areas of the search ference between intensification and
space with high quality solutions, and to diversification is that during an in-
move to unexplored areas of the search tensification stage the search focuses
space when necessary. The concepts for on examining neighbors of elite solu-
reaching these goals are nowadays called tions. [· · · ] The diversification stage
intensification and diversification. These on the other hand encourages the
terms stem from the TS field [Glover and search process to examine unvisited
Laguna 1997]. In other fields—such as the regions and to generate solutions
EC field—related concepts are often de- that differ in various significant ways
noted by exploitation (related to intensifi- from those seen before.” [Glover and
cation) and exploration (related to diversi- Laguna 1997]

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 293

Later in the same book, Glover and tion play balanced roles.” Yagiura and
Laguna write: “In some instances we may Ibaraki [2001].
conceive of intensification as having the
“Holland frames adaption as a ten-
function of an intermediate term strat-
sion between exploration (the search for
egy, while diversification applies to con-
new, useful adaptations) and exploitation
siderations that emerge in the longer
(the use and propagation of these adapta-
run.”
tions). The tension comes about since any
Furthermore, they write: “Strategic os- move toward exploration—testing previ-
cillation is closely linked to the origins ously unseen schemas or schemas whose
of tabu search, and provides a means instances seen so far have low fitness—
to achieve an effective interplay be- takes away from the exploitation of tried
tween intensification and diversifica- and true schemas. In any system (e.g., a
tion.” population of organisms) required to face
“After a local minimizer is encountered, environments with some degree of un-
all points in its attraction basin lose predictability, an optimal balance be-
any interest for optimization. The search tween exploration and exploitation
should avoid wasting excessive computing must be found. The system has to keep
time in a single basin and diversification trying out new possibilities (or else it could
should be activated. On the other hand, “over-adapt” and be inflexible in the face
in the assumptions that neighbors have of novelty), but it also has to continually
correlated cost function values, some ef- incorporate and use past experience as a
fort should be spent in searching for better guide for future behavior.”—M. Mitchell
points located close to the most recently citing J.H. Holland in Mitchell [1998].
found local minimum point (intensifica- All these descriptions share the com-
tion). The two requirements are con- mon view that there are two forces for
flicting and finding a proper balance which an appropriate balance has to be
of diversification and intensification found. Sometimes these two forces were
is a crucial issue in heuristics.” described as opposing forces. However,
[Battiti 1996]. lately some researchers raised the ques-
“A metaheuristic will be successful tion on how opposing intensification and
on a given optimization problem if it diversification really are.
can provide a balance between the ex- In 1998, Eiben and Schippers [1998]
ploitation of the accumulated search started a discussion about that in the field
experience and the exploration of the of Evolutionary Computation. They ques-
search space to identify regions with tion the common opinion about EC algo-
high quality solutions in a problem rithms, that they explore the search space
specific, near optimal way.” [Stützle by the genetic operators, while exploita-
1999b]. tion is achieved by selection. In their pa-
per they give examples of operators that
“Intensification is to search carefully one cannot unambiguously label as be-
and intensively around good solutions ing either intensification or diversifica-
found in the past search. Diversification, tion. So, for example, an operator using
on the contrary, is to guide the search a local search component to improve in-
to unvisited regions. These terminologies dividuals is not merely a mechanism of di-
are usually used to explain the basic ele- versification because it also comprises a
ments of Tabu Search, but these are essen- strong element of intensification (e.g., in
tial to all the metaheuristic algorithms. In Memetic Algorithms). Another example is
other words, various metaheuristic ideas the heuristically guided recombination of
should be understood from the viewpoint good quality solutions. If the use of the
of these two concepts, and metaheuris- accumulated search experience is identi-
tic algorithms should be designed so fied with intensification, then a recombi-
that intensification and diversifica- nation operator is not merely a means of

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


294 C. Blum and A. Roli

Fig. 18. The I&D frame provides a unified view on intensification and diversification in meta-
heuristics (OG = I&D components solely guided by the objective function, NOG = I&D components
solely guided by one or more function other than the objective function, R = I&D components solely
guided by randomness).

diversification, it also—as in the exam- ponent, and vice versa. To clarify this,
ple above—has a strong intensification we developed a framework to put I&D
component. components of different metaheuristics
Especially the TS literature advocates into relation with each other. We called
the view that intensification and diversifi- this framework—shown in Figure 18—the
cation cannot be characterized as opposing I&D frame.
forces. For example, in Glover and Laguna We depict the space of all I&D compo-
[1997], the authors write: “Similarly, as we nents as a triangle with the three corners
have noted, intensification and diversifica- corresponding to three extreme examples
tion are not opposed notions, for the best of I&D components. The corner denoted by
form of each contains aspects of the other, OG corresponds to I&D components solely
along a spectrum of alternatives.” guided by the objective function of the
Intensification and diversification can problem under consideration. An example
be considered as effects of algorithm of an I&D component which is located very
components. In order to understand sim- close to the corner OG is the steepest de-
ilarities and differences among meta- scent choice rule in local search. The cor-
heuristics, a framework may be helpful in ner denoted by NOG covers all I&D com-
providing a unified view on intensification ponents guided by one or more functions
and diversification components. We define other than the objective function, again
an I&D component as any algorithmic or without using any random component. An
functional component that has an inten- example for such a component is a de-
sification and/or a diversification effect terministic restart mechanism based on
on the search process. Accordingly, exam- global frequency counts of solution com-
ples of I&D components are genetic opera- ponents. The third corner, which is de-
tors, perturbations of probability distribu- noted by R, comprises all I&D compo-
tions, the use of tabu lists, or changes in nents that are completely random. This
the objective function. Thus, I&D compo- means that they are not guided by any-
nents are operators, actions, or strategies thing. For example, a restart of an EC
of metaheuristic algorithms. approach with random individuals is lo-
In contrast to the still widely spread cated in that corner. From the description
view that there are components that have of the corners, it becomes clear that cor-
either an intensification or a diversifica- ner OG corresponds to I&D components
tion effect, there are many I&D compo- with a maximum intensification effect and
nents that have both. In I&D components a minimum diversification effect. On the
that are commonly labelled as intensifi- other hand, corners NOG, R and the seg-
cation, the intensification component is ment between the two corners correspond
stronger than the diversification com- to I&D components with a maximum

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 295

diversification effect and a minimum in- ponents that are inherent to a metaheuris-
tensification effect.19 All I&D components tic and explain them in the context of the
can be located somewhere on or inbetween I&D frame. With that, we show that most
the three corners, where the intensifica- of the basic I&D components have an in-
tion effect is becoming smaller the further tensification character as well as a diver-
away a mechanism is located from OG. At sification character.
the same time the diversification effect is For many components and strategies of
growing. In step with this gradient is the metaheuristics, it is obvious that they in-
use I&D components make of the objec- volve an intensification as well as a diver-
tive function. The less an I&D component sification component, because they make
is using the objective function, the further an explicit use of the objective function.
away from corner OG it has to be located. For example, the basic idea of TS is a
There is also a second gradient to be found neighbor choice rule using one or more
in this frame (which is shown in the second tabu lists. This I&D component has two ef-
graphic of Figure 18). Corner R stands for fects on the search process. The restriction
complete randomness. The less random- of the set of possible neighbors in every
ness is involved in an I&D component, the step has a diversifying effect on the search,
further away from corner R it has to be lo- whereas the choice of the best neighbor
cated. Finally, a third gradient describes in the restricted set of neighbors (the best
the influence of criteria different from the non-tabu move) has an intensifying effect
objective function, which generally stem on the search. The balance between these
from the exploitation of the search history two effects can be varied by the length of
that is in some form kept in the memory. In the tabu list. Shorter tabu lists result in
the following, we analyze some basic I&D a lower influence of the diversifying ef-
components intrinsic to the basic versions fect, whereas longer tabu lists result in an
of the metaheuristics with respect to the overall higher influence of the diversifying
I&D frame. effect. The location of this component in
Figure 18 is on the segment between cor-
5.2. Basic I&D Components ner OG and NOG. The shorter the tabu
of Metaheuristics lists, the closer is the location to corner
OG, and vice versa.
The I&D components occurring in meta- Another example for such an I&D com-
heuristics can be divided in basic (or in- ponent is the probabilistic acceptance cri-
trinsic) ones and strategic ones. The basic terion in conjunction with the cooling
I&D components are the ones that are de- schedule in SA. The acceptance criterion
fined by the basic ideas of a metaheuris- is guided by the objective function and it
tic. On the other side, strategic I&D com- also involves a changing amount of ran-
ponents are composed of techniques and domness. The decrease of the tempera-
strategies the algorithm designer adds to ture parameter drives the system from di-
the basic metaheuristic in order to im- versification to intensification eventually
prove the performance by incorporating leading to a convergence20 of the system.
medium- and long-term strategies. Many Therefore, this I&D component is located
of these strategies were originally devel- in the interior of the I&D space between
oped in the context of a specific meta- corners OG, NOG and R.
heuristic. However, it becomes more and A third example is the following one.
more apparent that many of these strate- Ant Colony Optimization provides an I&D
gies can also be very useful when applied component that manages the update of the
in other metaheuristics. In the following, pheromone values. This component has
we exemplary choose some basic I&D com-

19 There is no quantitative difference between 20 Here, we use the term convergence in the sense
corners NOG and R. The difference is rather of getting stuck in the basin of attraction of a local
qualitative. minimum.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


296 C. Blum and A. Roli

the effect of changing the probability dis- one. Different landscapes differ in their
tribution that is used to sample the search ruggedness. A landscape with small (av-
space. It is guided by the objective func- erage) fitness differences between neigh-
tion (solution components found in better boring points is called smooth and it will
solutions than others are updated with usually have just a few local optima. In
a higher amount of pheromone) and it is contrast, a landscape with large (average)
also influenced by a function applying the fitness differences is called rugged and
pheromone evaporation. Therefore, this it will be usually characterized by many
component is located on the line between local optima. Most of the neighborhoods
corners OG and NOG. The effect of this used in metaheuristics provide some de-
mechanism is basically the intensification gree of smoothness that is higher than the
of the search, but there is also a diversify- one of a fitness landscape defined by a ran-
ing component that depends on the greed- dom neighborhood. This means that such
iness of the pheromone update (the less a neighborhood is, in a sense, preselecting
greedy or deterministic, the higher is the for every solution a set of neighbors for
diversifying effect). which the average fitness is not too dif-
For other strategies and components ferent. Therefore, even when a solution is
of metaheuristics, it is not immediately randomly selected from a set of neighbors,
obvious that they have both, an intensi- the objective function guidance is implic-
fication and a diversification effect. An itly present. The consequence is that even
example is the random selection of a for a random kick-move there is some de-
neighbor from the neighborhood of a cur- gree of intensification involved, as far as
rent solution, as it is done for example in a nonrandom neighborhood is considered.
the kick-move mechanism of ILS. Initially For a mutation operator of an EC
one might think that there is no intensifi- method that is doing a random change of
cation involved and that this mechanism a solution, it is neither immediately clear
has a pure diversification effect caused that it can have both, an intensification
by the use of randomness. However, for as well as a diversification effect. In the
the following reason, this is not the case. following, we assume a bit-string repre-
Many strategies (such as the kick-move sentation and a mutation operator that
operator mentioned above) involve the ex- is characterized by flipping every bit of
plicit or implicit use of a neighborhood. A a solution with a certain probability. The
neighborhood structures the search space implicit neighborhood used by this opera-
in the sense that it defines the topology tor is the completely connected neighbor-
of the so-called fitness landscape [Stadler hood. However, the neighbors have differ-
1995, 1996; Jones 1995a; Kauffman ent probabilities to be selected. The ones
1993], which can be visualized as a that are (with respect to the Hamming dis-
labelled graph. In this graph, nodes are tance) closer to the solution to which the
solutions (labels indicate their objective operator is applied to, have a higher proba-
function value) and arcs represent the bility to be generated by the operator. With
neighborhood relation between states.21 this observation, we can use the same ar-
A fitness landscape can be analyzed by gument as above in order to show an im-
means of statistical measures. One of the plicit use of objective function guidance.
common measures is the auto-correlation, The balance between intensification and
that provides information about how diversification is determined by the proba-
much the fitness will change when a move bility to flip each bit. The higher this prob-
is made from one point to a neighboring ability, the higher the diversification effect
of the operator. In contrast, the lower this
probability, the higher the intensification
21 The discussion of definitions and analysis of fitness
effect of this operator.
landscapes is beyond the scope of this article. We for-
ward the interested reader to Stadler [1995,1996];
On the other side, there are some strate-
Jones [1995a, 1995b]; Fonlupt et al. [1999], Hordijk gies that are often labelled as intensi-
[1996], Kauffman 1993], and Reeved [1999]. fication supposedly without having any

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 297

diversifying effect. One example is the se- Table 1. I&D-components intrinsic to the basic
lection operator in EC algorithms. How- metaheuristics
Metaheuristic I&D component
ever, nearly all selection operators involve
SA acceptance criterion
some degree of randomness (e.g., propor- + cooling schedule
tionate selection, tournament selection) TS neighbor choice (tabu lists)
and are therefore located somewhere be- aspiration criterion
tween corners OG, NOG and R of the I&D EC recombination
mutation
frame. This means that they also have a di- selection
versifying effect. The balance between in- ACO pheromone update
tensification and diversification depends probabilistic construction
on the function that assigns the selection ILS black box local search
probabilities. If the differences between kick-move
acceptance criterion
the selection probabilities are quite high, VNS black box local search
the intensification effect is higher, and neighborhood choice
similarly for the other extreme of having shaking phase
only small differences between the selec- acceptance criterion
tion probabilities. GRASP black box local search
restricted candidate list
Even an operator like the neighbor GLS penalty function
choice rule of a steepest descent local
search, which might be regarded as pure
intensification, has a diversifying compo- itly and explicitly, when strategies to guide
nent in the sense that the search is “mov- search algorithms are discussed.
ing” in the search space with respect to a The distinction between intensification
neighborhood. A neighborhood can be re- and diversification is often interpreted
garded as a function other than the objec- with respect to the temporal horizon of
tive function, making implicit use of the the search. Short-term search strategies
objective function. Therefore, a steepest can be seen as the iterative application of
descent local search is located between cor- tactics with a strong intensification char-
ners OG and NOG, and has both, a strong acter (for instance, the repeated applica-
intensification effect but also a weak di- tion of greedy moves). When the horizon
versification character. is enlarged, usually strategies referring
Based on these observations we con- to some sort of diversification come into
clude that probably most of the basic I&D play. Indeed, a general strategy usually
components used in metaheuristics have proves its effectiveness especially in the
both, an intensification and a diversifica- long term.
tion effect. However, the balance between The simplest strategy that coordinates
intensification and diversification might the interplay of intensification and diver-
be quite different for different I&D com- sification and can achieve an oscillating
ponents. Table 1 attempts to summarize balance between them is the restart mech-
the basic I&D components that are inher- anism: under certain circumstances (e.g.,
ent to the different metaheuristics. local optimum is reached, no improve-
ments after a specific number of algorithm
5.3. Strategic Control of Intensification
cycles, stagnation, no diversity) the algo-
and Diversification
rithm is restarted. The goal is to achieve
a sufficient coverage of the search space
The right balance between intensification in the long run, thus the already visited
and diversification is needed to obtain regions should not be explored again. The
an effective metaheuristic. Moreover, this computationally least expensive attempt
balance should not be fixed or only chang- to address this issue is a random restart.
ing into one direction (e.g., continuously Every algorithm applying this naive di-
increasing intensification). This balance versification mechanism therefore incor-
should rather be dynamical. This issue is porates an I&D component located in cor-
often treated in the literature, both implic- ner R of the I&D frame.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


298 C. Blum and A. Roli

Usually, the most effective restart ap- (ACS). This ACO algorithm uses an ad-
proaches make use of the search history. ditional I&D component aimed at intro-
Examples for such restart strategies are ducing diversification during the solution
the ones based on concepts such as global construction phase. While an ant is walk-
frequency and global desirability. The con- ing on the construction graph to construct
cept of global frequency is well known from a solution it reduces the pheromone val-
TS applications. In this concept, the num- ues on the nodes/arcs of the construction
ber of occurrences of solution components graph that it visits. This has the effect to
is counted during the run of the algorithm. reduce for the other ants the probability of
These numbers, called the global fre- taking the same path. This additional phe-
quency numbers, are then used for chang- romone update mechanism is called step-
ing the heuristic constructive method, for by-step online pheromone update rule. The
example to generate a new population for interplay between this component and the
restarting an EC method or the initial so- other pheromone update rules (online de-
lution for restarting a trajectory method. layed pheromone update rules and online
Similarly, the concept of global desirability pheromone update rule) leads to an oscil-
(which keeps for every solution component lating balance between intensification and
the objective function value of the best so- diversification.
lution it had been a member of) can be used Some more advanced strategies can be
to restart algorithms with a bias toward found in the literature. Often, they are
good quality solutions. I&D components described with respect to the particular
based on global frequency can be located metaheuristic in which they are applied.
in corner NOG, while global desirability- However, many of them are very general
based components are located along the and can be easily adapted and reused also
segment NOG-OG. Examples of the use in a different context. A very effective ex-
of nonrandom restarts can be found also ample is Strategic Oscillation [Glover and
in population-based methods. In EC algo- Laguna 1997].23 This strategy can be ap-
rithms, the new population can be gener- plied both to constructive methods and
ated by applying constructive heuristics22 improvement algorithms. Actions are in-
(line R-OG). In ACO, this goal is ad- voked with respect to a critical level (os-
dressed by smoothing or resetting phero- cillation boundary), which usually corre-
mone values [Stützle and Hoos 2000]. In sponds to a steady state of the algorithm.
the latter case, if the pheromone reset is Examples for steady states of an algorithm
also based on the search history, the ac- are local minima, completion of solution
tion is located inside the I&D frame. constructions, or the situation were no
There are also strategies explicitly components can be added to a partial solu-
aimed at dynamically changing the bal- tion such that it can be completed to a fea-
ance between intensification and diversi- sible solution. The oscillation strategy is
fication during the search. A fairly simple defined by a pattern indicating the way to
strategy is used in SA, where an increase approach the critical level, to cross it and
in diversification and simultaneous de- to cross it again from the other side. This
crease in intensification can be achieved pattern defines the distance of moves from
by “reheating” the system and then cool- the boundary and the duration of phases
ing it down again (which corresponds to (of intensification and diversification). Dif-
increasing parameter T and decreasing ferent patterns generate different strate-
it again according to some scheme). Such gies; moreover, they can also be adap-
a cooling scheme is called nonmonotonic tive and change depending on the current
cooling scheme (e.g., see Lundy and Mees state and history of the search process.
[1986] or Osman [1993]). Another exam- Other representative examples of general
ple can be found in Ant Colony System
23 Indeed, in Glover and Laguna [1997] and in the lit-
22 See,
for example, Freisleben and Merz [1996] and erature related to TS, many strategies are described
Grefenstette [1987]. and discussed.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 299

strategies that dynamically coordinate in- the successful applications that we have
tensification and diversification can be cited in previous sections are hybridiza-
found in Battiti and Protasi [1997] and tions. In the following, we distinguish dif-
Blum [2002a, 2002b]. ferent forms of hybridization. The first one
Furthermore, strategies are not re- consists of including components from one
stricted to single actions (e.g., variable as- metaheuristic into another one. The sec-
signments, moves), but may also guide ond form concerns systems that are some-
the application of coordinated sequences times labelled as cooperative search. They
of moves. Examples of such a strategy consist of various algorithms exchanging
are given by so-called ejection chain pro- information in some way. The third form
cedures [Glover and Laguna 1997; Rego is the integration of approximate and sys-
1998, 2001]. These procedures provide a tematic (or complete) methods. For a tax-
mechanism to perform compound moves, onomy of hybrid metaheuristics, see Talbi
that is, compositions of different types of [2002].
moves. For instance, in a problem defined Component Exchange Among Meta-
over a graph (e.g., the VRP), it is possi- heuristics. One of the most popular ways
ble to define two different moves: insertion of hybridization concerns the use of trajec-
and exchange of nodes; a compound move tory methods in population-based meth-
can thus be defined as the combination of ods. Most of the successful applications
an insertion and an exchange move. These of EC and ACO make use of local search
procedures describe general strategies to procedures. The reason for that becomes
combine the application of different neigh- apparent when analyzing the respec-
borhood structures, thus they provide tive strengths of trajectory methods and
an example of a general diversification/ population-based methods.
intensification interplay. Further exam- The power of population-based methods
ples of strategies that can be interpreted is certainly based on the concept of re-
as mechanisms to produce compositions combining solutions to obtain new ones.
of interlinked moves can also be found In EC algorithms and Scatter Search, ex-
in the literature concerning the integra- plicit recombinations are implemented by
tion of metaheuristics and complete tech- one or more recombination operators. In
niques [Caseau and Laburthe 1999; Shaw ACO and EDAs, recombination is implicit,
1998]. because new solutions are generated by
In conclusion, we would like to stress using a distribution over the search space
again that most metaheuristic compo- which is a function of earlier populations.
nents have both an intensification and a This allows to make guided steps in the
diversification effect. The higher the objec- search space, which are usually “larger”
tive function bias, the higher the intensi- than the steps done by trajectory meth-
fication effect. In contrast, diversification ods. In other words, a solution resulting
is achieved by following guiding criteria from a recombination in population-based
other than the objective function and also methods is usually more “different” from
by the use of randomness. With the in- the parents than, say, a predecessor solu-
troduction of the I&D frame, metaheuris- tion to a successor solution (obtained by
tics can be analyzed by their signature in applying a move) in TS. We also have “big”
the I&D frame. This can be a first step to- steps in trajectory methods like ILS and
ward the systematic design of metaheuris- VNS, but in these methods the steps are
tics, combining I&D components of differ- usually not guided (these steps are rather
ent origin. called “kick move” or “perturbation” indi-
cating the lack of guidance). It is interest-
ing to note, that in all population-based
5.4. Hybridization of Metaheuristics
methods there are mechanisms in which
We conclude our work by discussing a very good solutions found during the search
promising research issue: the hybridiza- influence the search process in the hope
tion of metaheuristics. In fact, many of to find better solutions in-between those

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


300 C. Blum and A. Roli

solutions and current solutions. In Path performed by possibly different algo-


Relinking, this idea is implemented in the rithms that exchange information about
most explicit way. The basic elements are states, models, entire sub-problems, so-
guiding solutions (which are the good solu- lutions or other search space character-
tions found) and initiating solutions. New istics. Typically, cooperative search algo-
solutions are produced by applying moves rithms consist of the parallel execution
to decrease the distance between the re- of search algorithms with a varying level
sulting solution and the guiding solution. of communication. The algorithms can be
In EC algorithms, this is often obtained by different or they can be instances of the
keeping the best (or a number of best) so- same algorithm working on different mod-
lution(s) found since the beginning of the els or running with different parameters
respective run of the algorithm in the po- setting. The algorithms composing a coop-
pulation. This is called a steady state evo- erative search system can be all approx-
lution process. Scatter Search performs a imate, all complete, or a mix of approxi-
steady state process by definition. In some mate and complete approaches.
ACO implementations (see, e.g., Stützle Cooperative search nowadays receives
and Hoos [2000] and Blum [2002b]) a phe- more attention, which is among other rea-
romone updating schedule is applied such sons due to the increasing research on
that in a situation where the algorithm parallel implementations of metaheuris-
has nearly converged to a solution, only tics. The aim of research on paralleliza-
the best found solution since the start of tion of metaheuristics is twofold. First,
the algorithm is used for updating the phe- metaheuristics should be redesigned to
romone trails. This corresponds to “chang- make them suitable for parallel imple-
ing direction” and directing the search pro- mentation in order to exploit intrinsic par-
cess toward a very good solution in the allelism. Second, an effective combination
hope to find better ones on the way. of metaheuristics has to be found, both
The strength of trajectory methods is to combine different characteristics and
rather to be found in the way they explore strengths and to design efficient commu-
a promising region in the search space. As nication mechanisms. Since the aim of this
in those methods local search is the driv- article is to provide an overview of the
ing component, a promising area in the core ideas and strategies of metaheuris-
search space is searched in a more struc- tics, we refer to Crainic and Toulouse
tured way than in population-based meth- [2002a, 2002b] for a survey on the state-
ods. In this way the danger of being close of-the-art in parallel metaheuristics.
to good solutions but “missing” them is not Integrating Metaheuristics and System-
as high as in population-based methods. atic Methods. Concluding this discus-
In summary, population-based methods sion on hybrid metaheuristics, we briefly
are better in identifying promising ar- overview the integration of metaheuristics
eas in the search space, whereas tra- and systematic search techniques. This
jectory methods are better in explor- approach has recently produced very ef-
ing promising areas in the search space. fective algorithms especially when applied
Thus, metaheuristic hybrids that in some to real-world problems. Discussions on
way manage to combine the advantage similarities, differences and possible in-
of population-based methods with the tegration of metaheuristics and system-
strength of trajectory methods are often atic search can be found in Freuder et al.
very successful. [1995], Ginsberg [1993], Harvey [1995],
Cooperative Search. A loose form of and Glover and Laguna [1997]. A very
hybridization is provided by cooperative successful example of such an integration
search [Hogg and Williams 1993; Hogg is the combination of metaheuristics and
and Huberman 1993; Bachelet and Talbi Constraint Programming (CP) [Focacci
2000; Denzinger and Offerman 1999; et al. 2002; Pesant and Gendreau 1996,
Toulouse et al. 1999a, 1999b; Sondergeld 1999; De Backer et al. 2000]. CP en-
and Voß 1999] which consists of a search ables to model a CO problem by means of

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 301

variables, domains24 and constraints, —The third possibility consists of intro-


which can be mathematical or symbolic ducing concepts or strategies from ei-
(global). The latter ones involve a set ther class of algorithms into the other.
of variables and describe subproblems, For example, the concepts of tabu list
thus reducing the modelling complexity and aspiration criteria—defined in Tabu
by encapsulating well defined parts of the Search—can be used to manage the list
problem into single constraints. Every of open nodes (i.e., the ones whose child
constraint is associated to a filtering al- nodes are not yet explored) in a tree
gorithm that deletes those values from a search algorithm.
variable domain that do not contribute
to feasible solutions. A CP system can The first approach can be seen as an in-
be seen as the interaction of components stance of cooperative search and it repre-
(constraints) which communicate through sents a rather loose integration.
shared variables. Constraints are acti- The second approach combines the ad-
vated as soon as a the domain of any vantages of a fast search space explo-
variable involved has been changed. Then, ration by means of a metaheuristic with
they perform a propagation phase, that is, the efficient neighborhood exploration per-
they apply the filtering algorithm. This be- formed by a systematic method. A promi-
havior stops as soon as there are no more nent example of such a kind of integra-
values that can be removed from the do- tion is Large Neighborhood Search and
mains or at least one domain is empty (i.e., related approaches [Shaw 1998; Caseau
no feasible solution exists). Since the com- and Laburthe 1999]. These approaches
plexity of the full constraint propagation are effective mainly when the neighbor-
is often exponential, the filtering is usu- hood to explore is very large. Moreover,
ally not complete. Therefore, at the end of many real-world problems have additional
the propagation phase, some domains may constraints (called side constraints) which
still contain unfeasible values. Hence, a might make them unsuitable for usual
search phase is started, such as Branch neighborhood exploration performed by
& Bound. A survey on the integration metaheuristics, since they usually just
of metaheuristics and CP is provided by sample the neighborhood or enumerate its
Focacci et al. [2002]. solutions. For instance, time window con-
There are three main approaches for the straints often reduce the number of fea-
integration of metaheuristics (especially sible solutions in a neighborhood, which
trajectory methods) and systematic tech- might make a local search inefficient.
niques (CP and tree search): Thus, domain filtering techniques can
effectively support neighborhood explo-
—A metaheuristic and a systematic ration. In fact, for such a kind of neigh-
method are sequentially applied (their borhoods, both sampling and enumeration
execution can be also interleaved). For are usually inefficient. More examples can
instance, the metaheuristic algorithm is be found in Pesant and Gendreau [1996,
run to produce some solutions that are 1999] and Focacci et al. [2002].
then used as heuristic information by The third approach preserves the search
the systematic search. Vice-versa, the space exploration based on a systematic
systematic algorithm can be run to gen- search (such as tree search), but sacri-
erate a partial solution which will then fices the exhaustive nature of the search
be completed by the metaheuristic. [Ginsberg 1993; Harvey 1995; Harvey and
—Metaheuristics use CP and/or tree Ginsberg 1995; Milano and Roli 2002].
search to efficiently explore the neigh- The hybridization is usually achieved by
borhood, instead of simply enumerating integrating concepts developed for meta-
the neighbors or randomly sampling the heuristics (e.g., probabilistic choices, as-
neighborhood. piration criteria, heuristic construction)
into tree search methods. A typical ap-
24 We restrict the discussion to finite domains. plication of this integration is the use

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


302 C. Blum and A. Roli

of a probabilistic backtracking, instead tasks” in the search process where some


of a—deterministic—chronological back- metaheuristics perform better than oth-
tracking. For instance, an extreme case is ers. This has to be examined more closely
the random choice of a backtracking move. in the future in order to be able to produce
The list of possible backtracking moves hybrid metaheuristics performing consid-
can also be sorted by means of a dynamic erably better than their “pure” parents. In
heuristic or a sampling of the leaves of fact we can find this phenomenon in many
the search tree. This sampling can be per- facets of life, not just in the world of al-
formed by a metaheuristic: the result of gorithms. Mixing and hybridizing is often
each possible backtracking move is chosen better than purity.
as the starting point for producing a com-
plete solution by a metaheuristic (more ACKNOWLEDGMENTS
than one solution can be generated from
each partial solution). Then, the quality We would like to thank Marco Dorigo, Joshua
of these complete solutions is used as Knowles, Andrea Lodi, Michela Milano, Michael
Sampels, and Thomas Stützle for suggestions and
a score for—probabilistically—selecting a
useful discussions. We also would like to thank the
backtracking move. Another prominent anonymous referees for many useful suggestions and
example is the introduction of randomiza- comments.
tion in systematic techniques, as described
in Gomes et al. [2000]. Many examples
REFERENCES
of this approach can be found in Focacci
et al. [2002], Jussien and Lhomme [2002], AARTS, E. H. L., KORST, J. H. M., AND LAARHOVEN,
Schaerf [1997], Dell’Amico and Lodi P. J. M. V. 1997. Simulated annealing. In
Local Search in Combinatorial Optimization,
[2002], Prestwich [2002], and Della Croce E. H. L. Aarts and J. K. Lenstra, Eds. Wiley-
and T’kindt [2003]. Interscience, Chichester, England, 91–120.
AARTS, E. H. L. AND LENSTRA, J. K., EDS. 1997.
Local Search in Combinatorial Optimization.
6. CONCLUSIONS Wiley, Chichester, UK.
In this work, we have presented and com- AIEX, R. M., BINATO, S., AND RESENDE, M. G. C. 2003.
Parallel GRASP with path-relinking for job shop
pared the current most important meta- scheduling. Paral. Comput. To appear.
heuristic methods. In Sections 3 and 4, BACHELET, V. AND TALBI, E. 2000. Cosearch: A
we have outlined the basic metaheuristics co-evolutionary metaheuritics. In Proceedings
as they are described in the literature. of Congress on Evolutionary Computation—
In Section 5, we then proposed a con- CEC’2000. 1550–1557.
ceptual comparison of the different meta- BÄCK, T. 1996. Evolutionary Algorithms in The-

heuristics based on the way they imple- ory and Practice. Oxford University Press, New
York.
ment the two main concepts for guiding
BÄCK, T., FOGEL, D. B., AND MACHALEWICZ, Z., EDS.
the search process: Intensification and di- 1997. Handbook of Evolutionary Computation.
versification. This comparison is founded Institute of Physics Publishing Ltd, Bristol, UK.
on the I&D frame, where algorithmic com- BALUJA, S. 1994. Population-based incremental
ponents can be characterized by the crite- learning: A method for integrating genetic
ria they depend upon (objective function, search based function optimization and compet-
itive learning. Tech. Rep. No. CMU-CS-94-163,
guiding functions and randomization) and Carnegie Mellon University, Pittsburgh, Pa.
their effect on the search process. Al- BALUJA, S. AND CARUANA, R. 1995. Removing the
though metaheuristics are different in the genetics from the standard genetic algorithm.
sense that some of them are population- In The International Conference on Machine
based (EC, ACO), and others are trajec- Learning 1995, A. Prieditis and S. Russel,
tory methods (SA, TS, ILS, VNS, GRASP), Eds. Morgan-Kaufmann Publishers, San Mateo,
Calif., 38–46.
and although they are based on differ-
BAR-YAM, Y. 1997. Dynamics of Complex Systems.
ent philosophies, the mechanisms to effi- Studies in nonlinearity. Addison–Wesley, Read-
ciently explore a search space are all based ing, Mass.
on intensification and diversification. Nev- BATTITI, R. 1996. Reactive search: Toward self-
ertheless, it is possible to identify “sub- tuning heuristics. In Modern Heuristic Search

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 303

Methods, V. J. Rayward-Smith, I. H. Osman, of GA-based multiobjective optimization tech-


C. R. Reeves, and G. D. Smith, Eds. Wiley, Chich- niques. ACM Comput. Surv. 32, 2, 109–143.
ester, UK, 61–83. CONNOLLY, D. T. 1990. An improved annealing
BATTITI, R. AND PROTASI, M. 1997. Reactive Search, scheme for the QAP. Europ. J. Oper. Res. 46, 93–
A history-base heuristic for MAX-SAT. ACM J. 100.
Exper. Algor. 2, Article 2. CRAINIC, T. G. AND TOULOUSE, M. 2002a. Intro-
BATTITI, R. AND TECCHIOLLI, G. 1994. The reactive duction to the special issue on parallel meta-
tabu search. ORSA J. Comput. 6, 2, 126–140. heuristics. J. Heuristics 8, 3, 247–249.
BINATO, S., HERY, W. J., LOEWENSTERN, D., AND RESENDE, CRAINIC, T. G. AND TOULOUSE, M. 2002b. Parallel
M. G. C. 2001. A greedy randomized adaptive strategies for meta-heuristics. In Handbook of
search procedure for job shop scheduling. In Es- Metaheuristics, F. Glover and G. Kochenberger,
says and Surveys on Metaheuristics, P. Hansen Eds. International Series in Operations Re-
and C. C. Ribeiro, Eds. Kluwer Academic search & Management Science, vol. 57. Kluwer
Publishers. Academic Publishers, Norwell, MA.
BLUM, C. 2002a. ACO applied to group shop sche- DE BACKER, B., FURNON, V., AND SHAW, P. 2000. Solv-
duling: A case study on intensification and di- ing Vehicle Routing Problems Using Constraint
versification. In Proceedings of ANTS 2002— Programming and Metaheuristics. J. Heuris-
Third International Workshop on Ant Algo- tics 6, 501–523.
rithms, M. Dorigo, G. Di Caro, and M. Sam- DE BONET, J. S., ISBELL JR., C. L., AND VIOLA, P. 1997.
pels, Eds. Lecture Notes in Computer Science, MIMIC: Finding optima by estimating probabil-
vol. 2463. Springer Verlag, Berlin, Germany, ity densities. In Proceedings of the 1997 Confer-
14–27. ence on Advances in Neural Information Process-
BLUM, C. 2002b. Ant colony optimization for the ing Systems (NIPS’97), M. C. Mozer, M. I. Jor-
edge-weighted k-cardinality tree problem. In dan, and T. Petsche, Eds. MIT Press, Cambridge,
GECCO 2002: Proceedings of the Genetic and MA, 424–431.
Evolutionary Computation Conference, W. B. DEJONG, K. A. 1975. An analysis of the behavior
Langdon, E. Cantú-Paz, K. Mathias, R. Roy, of a class of genetic adaptive systems. Ph.D. the-
D. Davis, R. Poli, K. Balakrishnan, V. Honavar, sis, University of Michigan, Ann Arbor, MI. Dis-
G. Rudolph, J. Wegener, L. Bull, M. A. Potter, sertation Abstracts International 36(10), 5140B,
A. C. Schultz, J. F. Miller, E. Burke, and University Microfilms Number 76-9381.
N. Jonoska, Eds. Morgan-Kaufmann, New York, DELLA CROCE, F. AND T’KINDT, V. 2003. A Recover-
27–34. ing Beam Search algorithm for the one machine
BLUM, C., ROLI, A., AND DORIGO, M. 2001. HC– dynamic total completion time scheduling prob-
ACO: The hyper-cube framework for ant colony lem. J. Oper. Res. Soc. To appear.
optimization. In Proceedings of MIC’2001— DELL’AMICO, M. AND LODI, A. 2002. On the inte-
Meta–heuristics International Conference. Vol. 2. gration of metaheuristic strategies in constraint
Porto, Portugal, 399–403. programming. In Adaptive Memory and Evolu-
CALÉGARY, P., CORAY, G., HERTZ, A., KOBLER, D., AND tion: Tabu Search and Scatter Search, C. Rego
KUONEN, P. 1999. A taxonomy of evolution- and B. Alidaee, Eds. Kluwer Academic Publish-
ary algorithms in combinatorial optimization. J. ers, Boston, MA.
Heuristics 5, 145–158. DELL’AMICO, M., LODI, A., AND MAFFIOLI, F. 1999. So-
CAMPOS, V., GLOVER, F., LAGUNA, M., AND MARTı́, R. lution of the cumulative assignment problem
2001. An experimental evaluation of a scatter with a well-structured tabu search method. J.
search for the linear ordering problem. J. Global Heuristics 5, 123–143.
Opt. 21, 397–414. DEN BESTEN, M. L., STÜTZLE, T., AND DORIGO, M. 2001.
CASEAU, Y. AND LABURTHE, F. 1999. Effective forget- Design of iterated local search algorithms: An
and-extend heuristics for scheduling problems. example application to the single machine total
In Proceedings of CP-AI-OR’02—Fourth Int. weighted tardiness problem. In Proceedings of
Workshop on Integration of AI and OR tech- EvoStim’01. Lecture Notes in Computer Science.
niques in Constraint Programming for Combi- Springer, 441–452.
natorial Optimization Problems. Ferrara (Italy). DENEUBOURG, J.-L., ARON, S., GOSS, S., AND PASTEELS, J.-
Also available at: www.deis.unibo.it/Events/ M. 1990. The self-organizing exploratory pat-
Deis/Workshops/Proceedings.html. tern of the argentine ant. J. Insect Behav. 3, 159–
CERNY, V. 1985. A thermodynamical approach to 168.
the travelling salesman problem: An efficient DENZINGER, J. AND OFFERMAN, T. 1999. On coopera-
simulation algorithm. J. Optim. Theory Appl. 45, tion between evolutionary algorithms and other
41–51. search paradigms. In Proceedings of Congress on
CHARDAIRE, P., LUTTON, J. L., AND SUTTER, A. 1995. Evolutionary Computation—CEC’1999. 2317–
Thermostatistical persistency: A powerful im- 2324.
proving concept for simulated annealing algo- DEVANEY, R. L. 1989. An introduction to chaotic
rithms. Europ. J. Oper. Res. 86, 565–579. dynamical systems, second ed. Addison–Wesley,
COELLO COELLO, C. A. 2000. An updated survey Reading, Mass.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


304 C. Blum and A. Roli

DI CARO, G. AND DORIGO, M. 1998. AntNet: Dis- application to industrial engineering problems.
tributed stigmergetic control for communication Comput. Indust. Eng. 37, 281–284.
networks. J. Artif. Int. Res. 9, 317–365. FLEISCHER, M. 1995. Simulated Annealing: past,
DORIGO, M. 1992. Optimization, learning and nat- present and future. In Proceedings of the 1995
ural algorithms (in italian). Ph.D. thesis, DEI, Winter Simulation Conference, C. Alexopoulos,
Politecnico di Milano, Italy. pp. 140. K. Kang, W. Lilegdon, and G. Goldsman, Eds.
DORIGO, M. AND DI CARO, G. 1999. The ant colony 155–161.
optimization meta-heuristic. In New Ideas FOCACCI, F., LABURTHE, F., AND LODI, A. 2002. Local
in Optimization, D. Corne, M. Dorigo, and search and constraint programming. In Hand-
F. Glover, Eds. McGraw-Hill, 11–32. book of Metaheuristics, F. Glover and G. Kochen-
DORIGO, M., DI CARO, G., AND GAMBARDELLA, L. M. berger, Eds. International Series in Opera-
1999. Ant algorithms for discrete optimization. tions Research & Management Science, vol. 57.
Art. Life 5, 2, 137–172. Kluwer Academic Publishers, Norwell, MA.
DORIGO, M. AND GAMBARDELLA, L. M. 1997. Ant FOGEL, D. B. 1994. An introduction to simulated
colony system: A cooperative learning approach evolutionary optimization. IEEE Trans. Neural
to the travelling salesman problem. IEEE Trans. Netw. 5, 1 (Jan.), 3–14.
Evolution. Comput. 1, 1 (Apr.), 53–66. FOGEL, G. B., PORTO, V. W., WEEKES, D. G., FOGEL, D. B.,
DORIGO, M., MANIEZZO, V., AND COLORNI, A. 1996. GRIFFEY, R. H., MCNEIL, J. A., LESNIK, E., ECKER,
Ant system: Optimization by a colony of cooper- D. J., AND SAMPATH, R. 2002. Discovery of RNA
ating agents. IEEE Trans. Syst. Man Cybernet.— structural elements using evolutionary compu-
Part B 26, 1, 29–41. tation. Nucleic Acids Res. 30, 23, 5310–5317.
DORIGO, M. AND STÜTZLE, T. 2002. The ant colony FOGEL, L. J. 1962. Toward inductive inference au-
optimization metaheuristic: Algorithms, appli- tomata. In Proceedings of the International Fed-
cations and advances. In Handbook of Meta- eration for Information Processing Congress.
heuristics, F. Glover and G. Kochenberger, Eds. Munich, 395–399.
International Series in Operations Research & FOGEL, L. J., OWENS, A. J., AND WALSH, M. J. 1966.
Management Science, vol. 57. Kluwer Academic Artificial Intelligence through Simulated Evolu-
Publishers, Norwell, MA, 251–285. tion. Wiley, New York.
DORIGO, M. AND STÜTZLE, T. 2003. Ant Colony Opti- FONLUPT, C., ROBILLIARD, D., PREUX, P., AND TALBI,
mization. MIT Press, Boston, MA. To appear. E. 1999. Fitness landscapes and performance
of meta-heuristics. In Meta-heuristics: advances
DUECK, G. 1993. New Optimization Heuristics. J.
and trends in local search paradigms for op-
Comput. Phy. 104, 86–92.
timization, S. Voß, S. Martello, I. Osman, and
DUECK, G. AND SCHEUER, T. 1990. Threshold accept- C. Roucairol, Eds. Kluwer Academic.
ing: A general purpose optimization algorithm
FREISLEBEN, B. AND MERZ, P. 1996. A genetic lo-
appearing superior to simulated annealing. J.
cal search algorithm for solving symmetric and
Comput. Phy. 90, 161–175.
asymmetric traveling salesman problems. In In-
EIBEN, A. E., RAUÉ, P.-E., AND RUTTKAY, Z. 1994. ternational Conference on Evolutionary Compu-
Genetic algorithms with multi-parent recombi- tation. 616–621.
nation. In Proceedings of the 3rd Conference
FREUDER, E. C., DECHTER, R., GINSBERG, M. L.,
on Parallel Problem Solving from Nature, Y.
SELMAN, B., AND TSANG, E. P. K. 1995. System-
Davidor, H.-P. Schwefel, and R. Manner, Eds.
atic versus stochastic constraint satisfaction. In
Lecture Notes in Computer Science, vol. 866.
Proceedings of the 14th International Joint Con-
Springer, Berlin, 78–87.
ference on Artificial Intelligence, IJCAI 1995.
EIBEN, A. E. AND RUTTKAY, Z. 1997. Constraint Vol. 2. Morgan-Kaufmann, 2027–2032.
satisfaction problems. In Handbook of Evolu- GAMBARDELLA, L. M. AND DORIGO, M. 2000. Ant
tionary Computation, T. Bäck, D. Fogel, and colony system hybridized with a new local search
M. Michalewicz, Eds. Institute of Physics Pub- for the sequential ordering problem. INFORMS
lishing Ltd, Bristol, UK. J. Comput. 12, 3, 237–255.
EIBEN, A. E. AND SCHIPPERS, C. A. 1998. On evo- GAREY, M. R. AND JOHNSON, D. S. 1979. Computers
lutionary exploration and exploitation. Fund. and Intractability; A Guide to the Theory of NP-
Inf. 35, 1–16. Completeness. W.H. Freeman.
FELLER, W. 1968. An Introduction to Probability GENDREAU, M., LAPORTE, G., AND POTVIN, J.-Y. 2001.
Theory and Its Applications. Wiley, New York. Metaheuristics for the vehicle routing problem.
FEO, T. A. AND RESENDE, M. G. C. 1995. Greedy ran- In The Vehicle Routing Problem, P. Toth and
domized adaptive search procedures. J. Global D. Vigo, Eds. SIAM Series on Discrete Mathe-
Optim. 6, 109–133. matics and Applications, vol. 9. 129–154.
FESTA, P. AND RESENDE, M. G. C. 2002. GRASP: An GINSBERG, M. L. 1993. Dynamic backtracking. J.
annotated bibliography. In Essays and Surveys Artif. Int. Res. 1, 25–46.
on Metaheuristics, C. C. Ribeiro and P. Hansen, GLOVER, F. 1977. Heuristics for integer program-
Eds. Kluwer Academic Publishers, 325–367. ming using surrogate constraints. Dec. Sci. 8,
FINK, A. AND VOß, S. 1999. Generic metaheuristics 156–166.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 305

GLOVER, F. 1986. Future paths for integer pro- HARIK, G. 1999. Linkage learning via probabilistic
gramming and links to artificial intelligence. modeling in the ECGA. Tech. Rep. No. 99010,
Comput. Oper. Res. 13, 533–549. IlliGAL, University of Illinois.
GLOVER, F. 1990. Tabu search Part II. ORSA J. HARVEY, W. D. 1995. Nonsystematic backtrack-
Comput. 2, 1, 4–32. ing search. Ph.D. thesis, CIRL, University of
GLOVER, F. 1999. Scatter search and path relink- Oregon.
ing. In New Ideas in Optimization, D. Corne, HARVEY, W. D. AND GINSBERG, M. L. 1995. Limited
M. Dorigo, and F. Glover, Eds. Advanced topics discrepancy search. In Proceedings of the 14th
in computer science series. McGraw-Hill. International Joint Conference on Artificial In-
GLOVER, F. AND LAGUNA, M. 1997. Tabu Search. telligence, IJCAI 1995 (Montréal, Qué, Canada).
Kluwer Academic Publishers. C. S. Mellish, Ed. Vol. 1. Morgan-Kaufmann,
607–615.
GLOVER, F., LAGUNA, M., AND MARTı́, R. 2000. Fun-
damentals of scatter search and path relinking. HERTZ, A. AND KOBLER, D. 2000. A framework
Control, 29, 3, 653–684. for the description of evolutionary algorithms.
Europ. J. Oper. Res. 126, 1–12.
GLOVER, F., LAGUNA, M., AND MARTı́, R. 2002. Scat-
ter search and path relinking: Advances and HOGG, T. AND HUBERMAN, A. 1993. Better than the
applications. In Handbook of Metaheuristics, best: The power of cooperation. In SFI 1992 Lec-
F. Glover and G. Kochenberger, Eds. Interna- tures in Complex Systems. Addison-Wesley, 163–
tional Series in Operations Research & Manage- 184.
ment Science, vol. 57. Kluwer Academic Publish- HOGG, T. AND WILLIAMS, C. 1993. Solving the really
ers, Norwell, MA. hard problems with cooperative search. In Pro-
GOLDBERG, D. E. 1989. Genetic Algorithms in ceedings of AAAI93. AAAI Press, 213–235.
Search, Optimization and Machine Learning. HOLLAND, J. H. 1975. Adaption in natural and artifi-
Addison Wesley, Reading, MA. cial systems. The University of Michigan Press,
GOLDBERG, D. E., DEB, K., AND KORB, B. 1991. Don’t Ann Harbor, MI.
worry, be messy. In Proceedings of the 4th In- HORDIJK, W. 1996. A measure of landscapes. Evo-
ternational Conference on Genetic Algorithms. lut. Comput. 4, 4, 335–360.
Morgan-Kaufmann, La Jolla, CA. INGBER, L. 1996. Adaptive simulated annealing
GOLDBERG, D. E. AND RICHARDSON, J. 1987. Genetic (ASA): Lessons learned. Cont. Cybernet.—
algorithms with sharing for multimodal function Special Issue on Simulated Annealing Applied
optimization. In Genetic Algorithms and their to Combinatorial Optimization 25, 1, 33–54.
Applications, J. J. Grefenstette, Ed. Lawrence JOHNSON, D. S. AND MCGEOCH, L. A. 1997. The trav-
Erlbaum Associates, Hillsdale, NJ, 41–49. eling salesman problem: a case study. In Local
GOMES, C. P., SELMAN, B., CRATO, N., AND KAUTZ, H. Search in Combinatorial Optimization, E. Aarts
2000. Heavy-Tayled phenomena in Satisfiabil- and J. Lenstra, Eds. Wiley, New York, 215–310.
ity and Constraint Satisfaction Prpblems. J. JONES, T. 1995a. Evolutionary algorithms, fitness
Automat. Reason. 24, 67–100. landscapes and search. Ph.D. thesis, Univ. of
GREFENSTETTE, J. J. 1987. Incorporating problem New Mexico, Albuquerque, NM.
specific knowledge into genetic algorithms. In JONES, T. 1995b. One operator, one landscape.
Genetic Algorithms and Simulated Annealing, Santa Fe Institute Tech. Rep. 95-02-025, Santa
L. Davis, Ed. Morgan-Kaufmann, 42–60. Fe Institute.
GREFENSTETTE, J. J. 1990. A user’s guide to GEN- JOSLIN, D. E. AND CLEMENTS, D. P. 1999. “Squeaky
ESIS 5.0. Tech. rep., Navy Centre for Applied Wheel” Optimization. J. Artif. Int. Res. 10, 353–
Research in Artificial Intelligence, Washington, 373.
D.C. JUSSIEN, N. AND LHOMME, O. 2002. Local search
HANSEN, P. 1986. The steepest ascent mildest de- with constraint propagation and conflict-based
scent heuristic for combinatorial programming. heuristics. Artif. Int. 139, 21–45.
In Congress on Numerical Methods in Combina- KAUFFMAN, S. A. 1993. The Origins of Order: Self-
torial Optimization. Capri, Italy. Organization and Selection in Evolution. Oxford
HANSEN, P. AND MLADENOVIĆ, N. 1997. Variable University Press.
neighborhood search for the p-median. Loc. KILBY, P., PROSSER, P., AND SHAW, P. 1999. Guided
Sci. 5, 207–226. Local Search for the Vehicle Routing Prob-
HANSEN, P. AND MLADENOVIĆ, N. 1999. An introduc- lem with time windows. In Meta-heuristics: Ad-
tion to variable neighborhood search. In Meta- vances and trends in local search paradigms for
heuristics: Advances and trends in local search optimization, S. Voß, S. Martello, I. Osman, and
paradigms for optimization, S. Voß, S. Martello, C. Roucairol, Eds. Kluwer Academic, 473–486.
I. Osman, and C. Roucairol, Eds. Kluwer Aca- KIRKPATRICK, S., GELATT, C. D., AND VECCHI, M. P.
demic Publishers, Chapter 30, 433–458. 1983. Optimization by simulated annealing.
HANSEN, P. AND MLADENOVIĆ, N. 2001. Variable Science, 13 May 1983 220, 4598, 671–680.
neighborhood search: Principles and applica- LAGUNA, M., LOURENÇO, H., AND MARTı́, R. 2000. As-
tions. Europ. J. Oper. Res. 130, 449–467. signing Proctors to Exams with Scatter Search.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


306 C. Blum and A. Roli

In Computing Tools for Modeling, Optimization lems. In SAT2000, I. Gent, H. van Maaren, and
and Simulation: Interfaces in Computer Science T. Walsh, Eds. IOS Press, 89–106.
and Operations Research, M. Laguna and J. L. MITCHELL, M. 1998. An Introduction to Genetic
González-Velarde, Eds. Kluwer Academic Pub- Algorithms. MIT press, Cambridge, MA.
lishers, Boston, MA, 215–227. MLADENOVIĆ, N. AND UROŠEVIĆ, D. 2001. Variable
LAGUNA, M. AND MARTı́, R. 1999. GRASP and path neighborhood search for the k-cardinality tree.
relinking for 2-layer straight line crossing mini- In Proceedings of MIC’2001—Meta–heuristics
mization. INFORMS J. Comput. 11, 1, 44–52. International Conference. Vol. 2. Porto, Portugal,
LAGUNA, M., MARTı́, R., AND CAMPOS, V. 1999. In- 743–747.
tensification and diversification with elite tabu MOSCATO, P. 1989. On evolution, search, optimiza-
search solutions for the linear ordering problem. tion, genetic algorithms and martial arts: To-
Comput. Oper. Res. 26, 1217–1230. ward memetic algorithms. Tech. Rep. Caltech
LARRAÑAGA, P. AND LOZANO, J. A., Eds. 2002. Es- Concurrent Computation Program 826, Califor-
timation of Distribution Algorithms: A New nia Institute of Technology,Pasadena, Calif.
Tool for Evolutionary Computation. Kluwer Aca- MOSCATO, P. 1999. Memetic algorithms: A short in-
demic Publishers, Boston, MA. troduction. In New Ideas in Optimization, F. G.
LOURENÇO, H. R., MARTIN, O., AND STÜTZLE, T. D. Corne and M. Dorigo, Eds. McGraw-Hill.
2001. A beginner’s introduction to Iterated MÜHLENBEIN, H. 1991. Evolution in time and
Local Search. In Proceedings of MIC’2001— space—The parallel genetic algorithm. In Foun-
Meta–heuristics International Conference. Vol. 1. dations of Genetic Algorithms, G. J. E. Rawlins,
Porto—Portugal, 1–6. Ed. Morgan-Kaufmann, San Mateo, Calif.
LOURENÇO, H. R., MARTIN, O., AND STÜTZLE, T. 2002. MÜHLENBEIN, H. AND PAAß, G. 1996. From recombi-
Iterated local search. In Handbook of Meta- nation of genes to the estimation of distributions.
heuristics, F. Glover and G. Kochenberger, Eds. In Proceedings of the 4th Conference on Paral-
International Series in Operations Research & lel Problem Solving from Nature—PPSN IV, H.-
Management Science, vol. 57. Kluwer Academic M. Voigt, W. Ebeling, I. Rechenberg, and H.-P.
Publishers, Norwell, MA, 321–353. Schwefel, Eds. Lecture Notes in Computer Sci-
LUNDY, M. AND MEES, A. 1986. Convergence of an ence, vol. 1411. Springer, Berlin, 178–187.
annealing algorithm. Math. Prog. 34, 1, 111–124. MÜHLENBEIN, H. AND VOIGT, H.-M. 1995. Gene pool
MARTIN, O. AND OTTO, S. W. 1996. Combining sim- recombination in genetic algorithms. In Proc. of
ulated annealing with local search heuristics. the Metaheuristics Conference, I. H. Osman and
Ann. Oper. Res. 63, 57–75. J. P. Kelly, Eds. Kluwer Academic Publishers,
MARTIN, O., OTTO, S. W., AND FELTEN, E. W. 1991. Norwell, USA.
Large-step Markov chains for the traveling NEMHAUSER, G. L. AND WOLSEY, A. L. 1988. Integer
salesman problem. Complex Syst. 5, 3, 299– and Combinatorial Optimization. Wiley, New
326. York.
MERKLE, D., MIDDENDORF, M., AND SCHMECK, H. NOWICKI, E. AND SMUTNICKI, C. 1996. A fast taboo
2002. Ant colony optimization for resource- search algorithm for the job-shop problem. Man-
constrained project scheduling. IEEE Trans. age. Sci. 42, 2, 797–813.
Evolut. Comput. 6, 4, 333–346. OSMAN, I. H. 1993. Metastrategy simulated an-
METAHEURISTICS NETWORK WEBSITE 2000. http://www. nealing and tabu search algorithms for the ve-
metaheuristics.net/. Visited in January 2003. hicle routing problem. Ann. Oper. Res. 41, 421–
MEULEAU, N. AND DORIGO, M. 2002. Ant colony opti- 451.
mization and stochastic gradient descent. Artif. OSMAN, I. H. AND LAPORTE, G. 1996. Metaheuristics:
Life 8, 2, 103–121. A bibliography. Ann. Oper. Res. 63, 513–623.
MICHALEWICZ, Z. AND MICHALEWICZ, M. 1997. Evolu- PAPADIMITRIOU, C. H. AND STEIGLITZ, K. 1982. Com-
tionary computation techniques and their ap- binatorial Optimization—Algorithms and Com-
plications. In Proceedings of the IEEE Inter- plexity. Dover Publications, Inc., New York.
national Conference on Intelligent Processing PELIKAN, M., GOLDBERG, D. E., AND CANTÚ-PAZ, E.
Systems, (Beijing, China). Institute of Elec- 1999a. BOA: The Bayesian optimization al-
trical & Electronics Engineers, Incorporated, gorithm. In Proceedings of the Genetic and
14–24. Evolutionary Computation Conference GECCO-
MILANO, M. AND ROLI, A. 2002. On the relation 99 (Orlando, Fla.). W. Banzhaf, J. Daida,
between complete and incomplete search: An A. E. Eiben, M. H. Garzon, V. Honavar,
informal discussion. In Proceedings of CP-AI- M. Jakiela, and R. E. Smith, Eds. Vol. I. Morgan-
OR’02—Fourth Int. Workshop on Integration of Kaufmann Publishers, San Fransisco, CA, 525–
AI and OR techniques in Constraint Program- 532.
ming for Combinatorial Optimization Problems PELIKAN, M., GOLDBERG, D. E., AND LOBO, F. 1999b.
(Le Croisic, France). 237–250. A survey of optimization by building and using
MILLS, P. AND TSANG, E. 2000. Guided local search probabilistic models. Tech. Rep. No. 99018, Illi-
for solving SAT and weighted MAX-SAT Prob- GAL, University of Illinois.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


Metaheuristics in Combinatorial Optimization 307

PESANT, G. AND GENDREAU, M. 1996. A view of local ternational Joint Conference on Artificial Intelli-
search in Constraint Programming. In Princi- gence, IJCAI 1997. Morgan-Kaufmann Publish-
ples and Practice of Constraint Programming— ers, San Mateo, CA, 1254–1259.
CP’96. Lecture Notes in Computer Science, vol. SCHAERF, A., CADOLI, M., AND LENZERINI, M. 2000.
1118. Springer-Verlag, 353–366. LOCAL++: a C++ framework for local search al-
PESANT, G. AND GENDREAU, M. 1999. A constraint gorithms. Softw. Pract. Exp. 30, 3, 233–256.
programming framework for local search meth- SHAW, P. 1998. Using constraint programming and
ods. J. Heuristics 5, 255–279. local search methods to solve vehicle routing
PITSOULIS, L. S. AND RESENDE, M. G. C. 2002. problems. In Principle and Practice of Constraint
Greedy randomized adaptive search proce- Programming—CP98, M. Maher and J.-F. Puget,
dure. In Handbook of Applied Optimization, P. Eds. Lecture Notes in Computer Science, vol.
Pardalos and M. Resende, Eds. Oxford Univer- 1520. Springer.
sity Press, 168–183. SIPPER, M., SANCHEZ, E., MANGE, D., TOMASSINI, M.,
PRAIS, M. AND RIBEIRO, C. C. 2000. Reactive PÉREZ-URIBE, A., AND STAUFFER, A. 1997. A
GRASP: An application to a matrix decompo- phylogenetic, ontogenetic, and epigenetic view
sition problem in TDMA traffic assignment. of bio-inspired hardware systems. IEEE Trans.
INFORMS J. Comput. 12, 164–176. Evolut. Comput. 1, 1, 83–97.
PRESTWICH, S. 2002. Combining the scalability of SONDERGELD, L. AND VOß, S. 1999. Cooperative in-
local search with the pruning techniques of telligent search using adaptive memory tech-
systematic search. Ann. Oper. Res. 115, 51– niques. In Meta-Heuristics: Advances and
72. Trends in Local Search Paradigms for Optimiza-
RADCLIFFE, N. J. 1991. Forma Analysis and Ran- tion, S. Voss, S. Martello, I. Osman, and C.
dom Respectful Recombination. In Proceed- Roucairol, Eds. Kluwer Academic Publishers,
ings of the Fourth International Conference Chapter 21, 297–312.
on Genetic Algorithms, ICGA 1991. Morgan- SPEARS, W. M., JONG, K. A. D., BÄCK, T., FOGEL, D. B.,
Kaufmann, San Mateo, Calif., 222–229. AND DE GARIS, H. 1993. An overview of evolu-
RAYWARD-SMITH, V. J. 1994. A unified approach to tionary computation. In Proceedings of the Euro-
tabu search, simulated annealing and genetic al- pean Conference on Machine Learning (ECML-
gorithms. In Applications of Modern Heuristics, 93), P. B. Brazdil, Ed. Vol. 667. Springer Verlag,
V. J. Rayward-Smith, Ed. Alfred Waller Limited, Vienna, Austria, 442–459.
Publishers. STADLER, P. F. 1995. Towards a theory of land-
RECHENBERG, I. 1973. Evolutionsstrategie: Opti- scapes. In Complex Systems and Binary Net-
mierung technischer Systeme nach Prinzip- works, R. Lopéz-Peña, R. Capovilla, R. Garcı́a-
ien der biologischen Evolution. Frommann- Pelayo, H. Waelbroeck, and F. Zertuche, Eds.
Holzboog. Lecture Notes in Physics, vol. 461. Springer-
REEVES, C. R., Ed. 1993. Modern Heuristic Tech- Verlag, Berlin, New York, 77–163. Also available
niques for Combinatorial Problems. Blackwell as SFI preprint 95-03-030.
Scientific Publishing, Oxford, England. STADLER, P. F. 1996. Landscapes and their corre-
REEVES, C. R. 1999. Landscapes, operators and lation functions. J. Math. Chem. 20, 1–45. Also
heuristic search. Ann. Oper. Res. 86, 473–490. available as SFI preprint 95-07-067.
REEVES, C. R. AND ROWE, J. E. 2002. Genetic Al- STÜTZLE, T. 1999a. Iterated local search for the
gorithms: Principles and Perspectives. A Guide quadratic assignment problem. Tech. rep. aida-
to GA Theory. Kluwer Academic Publishers, 99-03, FG Intellektik, TU Darmstadt.
Boston (USA). STÜTZLE, T. 1999b. Local Search Algorithms for
REGO, C. 1998. Relaxed Tours and Path Ejections Combinatorial Problems—Analysis, Algorithms
for the Traveling Salesman Problem. Europ. J. and New Applications. DISKI—Dissertationen
Oper. Res. 106, 522–538. zur Künstliken Intelligenz. infix, Sankt
REGO, C. 2001. Node-ejection chains for the ve- Augustin, Germany.
hicle routing problem: Sequential and paral- STÜTZLE, T. AND HOOS, H. H. 2000. MAX -MI N
lel algorithms. Paral. Comput. 27, 3, 201– Ant System. Fut. Gen. Comput. Syst. 16, 8, 889–
222. 914.
RESENDE, M. G. C. AND RIBEIRO, C. C. 1997. A SYSWERDA, G. 1993. Simulated Crossover in Ge-
GRASP for graph planarization. Networks 29, netic Algorithms. In Proceedings of the 2nd
173–189. Workshop on Foundations of Genetic Algorithms,
RIBEIRO, C. C. AND SOUZA, M. C. 2002. Variable L. Whitley, Ed. Morgan-Kaufmann Publishers,
neighborhood search for the degree constrained San Mateo, Calif., 239–255.
minimum spanning tree problem. Disc. Appl. TABU SEARCH WEBSITE. 2003. http://www.tabusearch.
Math. 118, 43–54. net. Visited in January 2003.
SCHAERF, A. 1997. Combining local search and TAILLARD, E. 1991. Robust Taboo Search for the
look-ahead for scheduling and constraint satis- Quadratic Assignment Problem. Paral. Com-
faction problems. In Proceedings of the 15th In- put. 17, 443–455.

ACM Computing Surveys, Vol. 35, No. 3, September 2003.


308 C. Blum and A. Roli

TALBI, E.-G. 2002. A Taxonomy of Hybrid Meta- VOß, S. AND WOODRUFF, D., Eds. 2002. Optimization
heuristics. Journal of Heuristics 8, 5, 541–564. Software Class Libraries. Kluwer Academic Pub-
TOULOUSE, M., CRAINIC, T., AND SANSÒ, B. 1999a. lishers, Dordrecht, The Netherlands.
An experimental study of the systemic behav- VOUDOURIS, C. 1997. Guided local search for com-
ior of cooperative search algorithms. In Meta- binatorial optimization problems. Ph.D. disser-
Heuristics: Advances and Trends in Local Search tation, Department of Computer Science, Uni-
Paradigms for Optimization, S. Voß, S. Martello, versity of Essex. pp. 166.
I. Osman, and C. Roucairol, Eds. Kluwer Aca- VOUDOURIS, C. AND TSANG, E. 1999. Guided local
demic Publishers, Chapter 26, 373–392. search. Europ. J. Oper. Res. 113, 2, 469–499.
TOULOUSE, M., THULASIRAMAN, K., AND GLOVER, F. WADE, A. S. AND RAYWARD-SMITH, V. J. 1997. Ef-
1999b. Multi-level cooperative search: A new fective local search for the steiner tree prob-
paradigm for combinatorial optimization and ap- lem. Studies in Locational Analysis 11, 219–
plication to graph partitioning. In Proceedings 241. Also in Advances in Steiner Trees, ed. by
of the 5th International Euro-Par Conference Ding-Zhu Du, J. M.Smith and J.H. Rubinstein,
on Parallel Processing. Lecture Notes in Com- Kluwer, 2000.
puter Science. Springer-Verlag, New York, 533– WATSON, R. A., HORNBY, G. S., AND POLLACK, J. B. 1998.
542.
Modeling building-block interdependency. In
VAN KEMENADE, C. H. M. 1996. Explicit filtering Late Breaking Papers at the Genetic Program-
of building blocks for genetic algorithms. In ming 1998 Conference, J. R. Koza, Ed. Stanford
Proceedings of the 4th Conference on Parallel University Bookstore, University of Wisconsin,
Problem Solving from Nature—PPSN IV, H.- Madison, Wisconsin, USA.
M. Voigt, W. Ebeling, I. Rechenberg, and H.-P.
WHITLEY, D. 1989. The GENITOR algorithm and
Schwefel, Eds. Lecture Notes in Computer Sci- selective pressure: Why rank-based allocation
ence, vol. 1141. Springer, Berlin, 494–503. of reproductive trials is best. In Proceedings of
VAN LAARHOVEN, P. J. M., AARTS, E. H. L., AND LENSTRA, the 3rd International Conference on Genetic Al-
J. K. 1992. Job Shop Scheduling by Simulated gorithms, ICGA 1989. Morgan-Kaufmann Pub-
Annealing. Oper. Res. 40, 113–125. lishers, 116–121.
VOSE, M. 1999. The Simple Genetic Algorithm: YAGIURA, M. AND IBARAKI, T. 2001. On metaheuris-
Foundations and Theory. Complex Adaptive Sys- tic algorithms for combinatorial optimization
tems. MIT Press. problems. Syst. Comput. Japan 32, 3, 33–
VOß, S., MARTELLO, S., OSMAN, I. H., AND ROUCAIROL, 55.
C., Eds. 1999. Meta-Heuristics—Advances ZLOCHIN, M., BIRATTARI, M., MEULEAU, N., AND DORIGO,
and Trends in Local Search Paradigms for M. 2004. Model-based search for combinato-
Optimization. Kluwer Academic Publishers, rial optimization: A critical survey. Ann. Oper.
Dordrecht, The Netherlands. Res. To appear.

Received July 2002; revised February 2003; accepted June 2003

ACM Computing Surveys, Vol. 35, No. 3, September 2003.

Das könnte Ihnen auch gefallen