Results 11  20
of
217
Variable neighborhood search: Principles and applications
, 2001
"... Systematic change of neighborhood within a possibly randomized local search algorithm yields a simple and effective metaheuristic for combinatorial and global optimization, called variable neighborhood search (VNS). We present a basic scheme for this purpose, which can easily be implemented using an ..."
Abstract

Cited by 94 (9 self)
 Add to MetaCart
Systematic change of neighborhood within a possibly randomized local search algorithm yields a simple and effective metaheuristic for combinatorial and global optimization, called variable neighborhood search (VNS). We present a basic scheme for this purpose, which can easily be implemented using any local search algorithm as a subroutine. Its effectiveness is illustrated by solving several classical combinatorial or global optimization problems. Moreover, several extensions are proposed for solving large problem instances: using VNS within the successive approximation method yields a twolevel VNS, called variable neighborhood decomposition search (VNDS); modifying the basic scheme to explore easily valleys far from the incumbent solution yields an efficient skewed VNS (SVNS) heuristic. Finally, we show how to stabilize column generation algorithms with help of VNS and discuss various ways to use VNS in graph theory, i.e., to suggest, disprove or give hints on how to prove conjectures, an area where metaheuristics do not appear
Nearly Linear Time Approximation Schemes for Euclidean TSP and other Geometric Problems
, 1997
"... We present a randomized polynomial time approximation scheme for Euclidean TSP in ! 2 that is substantially more efficient than our earlier scheme in [2] (and the scheme of Mitchell [21]). For any fixed c ? 1 and any set of n nodes in the plane, the new scheme finds a (1+ 1 c )approximation to ..."
Abstract

Cited by 91 (4 self)
 Add to MetaCart
We present a randomized polynomial time approximation scheme for Euclidean TSP in ! 2 that is substantially more efficient than our earlier scheme in [2] (and the scheme of Mitchell [21]). For any fixed c ? 1 and any set of n nodes in the plane, the new scheme finds a (1+ 1 c )approximation to the optimum traveling salesman tour in O(n(log n) O(c) ) time. (Our earlier scheme ran in n O(c) time.) For points in ! d the algorithm runs in O(n(log n) (O( p dc)) d\Gamma1 ) time. This time is polynomial (actually nearly linear) for every fixed c; d. Designing such a polynomialtime algorithm was an open problem (our earlier algorithm in [2] ran in superpolynomial time for d 3). The algorithm generalizes to the same set of Euclidean problems handled by the previous algorithm, including Steiner Tree, kTSP, kMST, etc, although for kTSP and kMST the running time gets multiplied by k. We also use our ideas to design nearlylinear time approximation schemes for Euclidean vers...
MAXMIN Ant System
 FUTURE GENERATION COMPUTER SYSTEMS
, 2000
"... Ant System, the first Ant Colony Optimization algorithm, showed to be a viable method for attacking hard combinatorial optimization problems. Yet, its performance, when compared to more finetuned algorithms, was rather poor for large instances of traditional benchmark problems like the Traveling Sa ..."
Abstract

Cited by 79 (4 self)
 Add to MetaCart
Ant System, the first Ant Colony Optimization algorithm, showed to be a viable method for attacking hard combinatorial optimization problems. Yet, its performance, when compared to more finetuned algorithms, was rather poor for large instances of traditional benchmark problems like the Traveling Salesman Problem. To show that Ant Colony Optimization algorithms could be good alternatives to existing algorithms for hard combinatorial optimization problems, recent research in this ares has mainly focused on the development of algorithmic variants which achieve better performance than AS. In this article, we present ¨�©� � –¨��� � Ant System, an Ant Colony Optimization algorithm derived from Ant System. ¨�©� � –¨��� � Ant System differs from Ant System in several important aspects, whose usefulness we demonstrate by means of an experimental study. Additionally, we relate one of the characteristics specific to ¨� ¨ AS — that of using a greedier search than Ant System — to results from the search space analysis of the combinatorial optimization problems attacked in this paper. Our computational results on the Traveling Salesman Problem and the Quadratic Assignment Problem show that ¨�©� � – ¨��� � Ant System is currently among the best performing algorithms for these problems.
A Theoretician's Guide to the Experimental Analysis of Algorithms
, 1996
"... This paper presents an informal discussion of issues that arise when one attempts to analyze algorithms experimentally. It is based on lessons learned by the author over the course of more than a decade of experimentation, survey paper writing, refereeing, and lively discussions with other experimen ..."
Abstract

Cited by 77 (0 self)
 Add to MetaCart
This paper presents an informal discussion of issues that arise when one attempts to analyze algorithms experimentally. It is based on lessons learned by the author over the course of more than a decade of experimentation, survey paper writing, refereeing, and lively discussions with other experimentalists. Although written from the perspective of a theoretical computer scientist, it is intended to be of use to researchers from all fields who want to study algorithms experimentally. It has two goals: first, to provide a useful guide to new experimentalists about how such work can best be performed and written up, and second, to challenge current researchers to think about whether their own work might be improved from a scientific point of view. With the latter purpose in mind, the author hopes that at least a few of his recommendations will be considered controversial.
The ant colony optimization metaheuristic: Algorithms, applications, and advances
 Handbook of Metaheuristics
, 2002
"... ..."
Genetic Local Search for the TSP: New Results
 In Proceedings of the 1997 IEEE International Conference on Evolutionary Computation
, 1997
"... The combination of local search heuristics and genetic algorithms has been shown to be an effective approach for finding nearoptimum solutions to the traveling salesman problem. In this paper, previously proposed genetic local search algorithms for the symmetric and asymmetric traveling salesman pr ..."
Abstract

Cited by 74 (13 self)
 Add to MetaCart
The combination of local search heuristics and genetic algorithms has been shown to be an effective approach for finding nearoptimum solutions to the traveling salesman problem. In this paper, previously proposed genetic local search algorithms for the symmetric and asymmetric traveling salesman problem are revisited and potential improvements are identified. Since local search is the central component in which most of the computation time is spent, improving the efficiency of the local search operators is crucial for improving the overall performance of the algorithms. The modifications of the algorithms are described and the new results obtained are presented. The results indicate that the improved algorithms are able to arrive at better solutions in significantly less time. I. Introduction Consider a salesman who wants to start from his home city, visit each of a set of n cities exactly once, and then return home. Since the salesman is interested in finding the shortest possible r...
Parameterized Complexity: A Framework for Systematically Confronting Computational Intractability
 DIMACS Series in Discrete Mathematics and Theoretical Computer Science
, 1997
"... In this paper we give a programmatic overview of parameterized computational complexity in the broad context of the problem of coping with computational intractability. We give some examples of how fixedparameter tractability techniques can deliver practical algorithms in two different ways: (1) by ..."
Abstract

Cited by 72 (15 self)
 Add to MetaCart
In this paper we give a programmatic overview of parameterized computational complexity in the broad context of the problem of coping with computational intractability. We give some examples of how fixedparameter tractability techniques can deliver practical algorithms in two different ways: (1) by providing useful exact algorithms for small parameter ranges, and (2) by providing guidance in the design of heuristic algorithms. In particular, we describe an improved FPT kernelization algorithm for Vertex Cover, a practical FPT algorithm for the Maximum Agreement Subtree (MAST) problem parameterized by the number of species to be deleted, and new general heuristics for these problems based on FPT techniques. In the course of making this overview, we also investigate some structural and hardness issues. We prove that an important naturally parameterized problem in artificial intelligence, STRIPS Planning (where the parameter is the size of the plan) is complete for W [1]. As a corollary, this implies that kStep Reachability for Petri Nets is complete for W [1]. We describe how the concept of treewidth can be applied to STRIPS Planning and other problems of logic to obtain FPT results. We describe a surprising structural result concerning the top end of the parameterized complexity hierarchy: the naturally parameterized Graph kColoring problem cannot be resolved with respect to XP either by showing membership in XP, or by showing hardness for XP without settling the P = NP question one way or the other.
Solving Symmetric and Asymmetric TSPs by Ant Colonies
, 1996
"... In this paper we present ACS, a distributed algorithm for the solution of combinatorial optimization problems which was inspired by the observation of real colonies of ants. We apply ACS to both symmetric and asymmetric traveling salesman problems. Results show that ACS is able to find good sol ..."
Abstract

Cited by 62 (17 self)
 Add to MetaCart
In this paper we present ACS, a distributed algorithm for the solution of combinatorial optimization problems which was inspired by the observation of real colonies of ants. We apply ACS to both symmetric and asymmetric traveling salesman problems. Results show that ACS is able to find good solutions to these problems. I. Introduction In this paper we present Ant Colony System (ACS), a novel distributed approach to combinatorial optimization based on the observation of real ant colonies behavior. ACS finds its ground in one of the authors previous work on the socalled Ant System (AS) [1],[2],[5],[7] and in AntQ [8] an extension of AS with Qlearning [12], a reinforcement learning technique. In particular, ACS is a revisited version of AntQ where a different way to update the experience accumulated by the artificial ants has been introduced [6]. All the mentioned systems belong to the Artificial Ant Colonies (AAC) family of algorithms that has been applied to various combinat...
Fitness Landscapes and Memetic Algorithm Design
 New Ideas in Optimization
, 1999
"... Introduction The notion of fitness landscapes has been introduced to describe the dynamics of evolutionary adaptation in nature [40] and has become a powerful concept in evolutionary theory. Fitness landscapes are equally well suited to describe the behavior of heuristic search methods in optimizat ..."
Abstract

Cited by 58 (7 self)
 Add to MetaCart
Introduction The notion of fitness landscapes has been introduced to describe the dynamics of evolutionary adaptation in nature [40] and has become a powerful concept in evolutionary theory. Fitness landscapes are equally well suited to describe the behavior of heuristic search methods in optimization, since the process of evolution can be thought of as searching a collection of genotypes in order to find the genotype of an organism with highest fitness and thus highest chance of survival. Thinking of a heuristic search method as a strategy to "navigate" in the fitness landscape of a given optimization problem may help in predicting the performance of a heuristic search algorithm if the structure of the landscape is known in advance. Furthermore, the analysis of fitness landscapes may help in designing highly effective search algorithms. In the following we show how the analysis of fitness landscapes of combinatorial optimization problems can aid in designing the components of
Learning Evaluation Functions to Improve Optimization by Local Search
 Journal of Machine Learning Research
, 2000
"... This paper describes algorithms that learn to improve search performance on largescale optimization tasks. The main algorithm, Stage, works by learning an evaluation function that predicts the outcome of a local search algorithm, such as hillclimbing or Walksat, from features of states visited durin ..."
Abstract

Cited by 56 (0 self)
 Add to MetaCart
This paper describes algorithms that learn to improve search performance on largescale optimization tasks. The main algorithm, Stage, works by learning an evaluation function that predicts the outcome of a local search algorithm, such as hillclimbing or Walksat, from features of states visited during search. The learned evaluation function is then used to bias future search trajectories toward better optima on the same problem. Another algorithm, XStage, transfers previously learned evaluation functions to new, similar optimization problems. Empirical results are provided on seven largescale optimization domains: binpacking, channel routing, Bayesian network structurefinding, radiotherapy treatment planning, cartogram design, Boolean satisfiability, and Boggle board setup.