Results 1  10
of
82
A comparative analysis of selection schemes used in genetic algorithms
 Foundations of Genetic Algorithms
, 1991
"... This paper considers a number of selection schemes commonly used in modern genetic algorithms. Specifically, proportionate reproduction, ranking selection, tournament selection, and Genitor (or «steady state") selection are compared on the basis of solutions to deterministic difference or diffe ..."
Abstract

Cited by 392 (32 self)
 Add to MetaCart
This paper considers a number of selection schemes commonly used in modern genetic algorithms. Specifically, proportionate reproduction, ranking selection, tournament selection, and Genitor (or «steady state") selection are compared on the basis of solutions to deterministic difference or differential equations, which are verified through computer simulations. The analysis provides convenient approximate or exact solutions as well as useful convergence time and growth ratio estimates. The paper recommends practical application of the analyses and suggests a number of paths for more detailed analytical investigation of selection techniques. Keywords: proportionate selection, ranking selection, tournament selection, Genitor, takeover time, time complexity, growth ratio. 1
An Overview of Evolutionary Algorithms in Multiobjective Optimization
 Evolutionary Computation
, 1995
"... The application of evolutionary algorithms (EAs) in multiobjective optimization is currently receiving growing interest from researchers with various backgrounds. Most research in this area has understandably concentrated on the selection stage of EAs, due to the need to integrate vectorial performa ..."
Abstract

Cited by 361 (10 self)
 Add to MetaCart
The application of evolutionary algorithms (EAs) in multiobjective optimization is currently receiving growing interest from researchers with various backgrounds. Most research in this area has understandably concentrated on the selection stage of EAs, due to the need to integrate vectorial performance measures with the inherently scalar way in which EAs reward individual performance, i.e., number of offspring. In this review, current multiobjective evolutionary approaches are discussed, ranging from the conventional analytical aggregation of the different objectives into a single function to a number of populationbased approaches and the more recent ranking schemes based on the definition of Paretooptimality. The sensitivity of different methods to
Genetic Algorithms, Noise, and the Sizing of Populations
 COMPLEX SYSTEMS
, 1991
"... This paper considers the effect of stochasticity on the quality of convergence of genetic algorithms (GAs). In many problems, the variance of buildingblock fitness or socalled collateral noise is the major source of variance, and a populationsizing equation is derived to ensure that average sig ..."
Abstract

Cited by 240 (85 self)
 Add to MetaCart
This paper considers the effect of stochasticity on the quality of convergence of genetic algorithms (GAs). In many problems, the variance of buildingblock fitness or socalled collateral noise is the major source of variance, and a populationsizing equation is derived to ensure that average signaltocollateralnoise ratios are favorable to the discrimination of the best building blocks required to solve a problem of bounded deception. The sizing relation is modified to permit the inclusion of other sources of stochasticity, such as the noise of selection, the noise of genetic operators, and the explicit noise or nondeterminism of the objective function. In a test suite of five functions, the sizing relation proves to be a conservative predictor of average correct convergence, as long as all major sources of noise are considered in the sizing calculation. These results suggest how the sizing equation may be viewed as a coarse delineation of a boundary between what a physicist might call two distinct phases of GA behavior. At low population sizes the GA makes many errors of decision, and the quality of convergence is largely left to the vagaries of chance or the serial fixup of flawed results through mutation or other serial injection of diversity. At large population sizes, GAs can reliably discriminate between good and bad building blocks, and parallel processing and recombination of building blocks lead to quick solution of even difficult deceptive problems. Additionally, the paper outlines a number of extensions to this work, including the development of more refined models of the relation between generational average error and ultimate convergence quality, the development of online methods for sizing populations via the estimation of populations...
Niching Methods for Genetic Algorithms
, 1995
"... Niching methods extend genetic algorithms to domains that require the location and maintenance of multiple solutions. Such domains include classification and machine learning, multimodal function optimization, multiobjective function optimization, and simulation of complex and adaptive systems. This ..."
Abstract

Cited by 193 (1 self)
 Add to MetaCart
Niching methods extend genetic algorithms to domains that require the location and maintenance of multiple solutions. Such domains include classification and machine learning, multimodal function optimization, multiobjective function optimization, and simulation of complex and adaptive systems. This study presents a comprehensive treatment of niching methods and the related topic of population diversity. Its purpose is to analyze existing niching methods and to design improved niching methods. To achieve this purpose, it first develops a general framework for the modelling of niching methods, and then applies this framework to construct models of individual niching methods, specifically crowding and sharing methods. Using a constructed model of crowding, this study determines why crowding methods over the last two decades have not made effective niching methods. A series of tests and design modifications results in the development of a highly effective form of crowding, called determin...
Searching for Diverse, Cooperative Populations with Genetic Algorithms
 EVOLUTIONARY COMPUTATION
, 1993
"... In typical applications, genetic algorithms (GAs) process populations of potential problem solutions to evolve a single population member that specifies an "optimized" solution. The majority of GA analysis has focused on these optimization applications. In other applications (notably learning cla ..."
Abstract

Cited by 94 (10 self)
 Add to MetaCart
In typical applications, genetic algorithms (GAs) process populations of potential problem solutions to evolve a single population member that specifies an "optimized" solution. The majority of GA analysis has focused on these optimization applications. In other applications (notably learning classifier systems and certain connectionist learning systems), a GA searches for a population of cooperative structures that jointly perform a computational task. This paper presents an analysis of this type of GA problem. The analysis considers a simplified geneticsbased machine learning system: a model of an immune system. In this model, a GA must discover a set of patternmatching antibodies that effectively match a set of antigen patterns. Analysis shows how a GA can automatically evolve and sustain a diverse, cooperative population. The cooperation emerges as a natural part of the antigenantibody matching procedure. This emergent effect is shown to be similar to fitness sharing, ...
An Overview of Genetic Algorithms: Part 1, Fundamentals
, 1993
"... this article may be reproduced for commercial purposes. 1 Introduction ..."
Abstract

Cited by 81 (1 self)
 Add to MetaCart
this article may be reproduced for commercial purposes. 1 Introduction
SearchIntensive Concept Induction
, 1995
"... This paper describes REGAL, a distributed genetic algorithmbased system, designed for learning First Order Logic concept descriptions from examples. The system is a hybrid between the Pittsburgh and the Michigan approaches, as the population constitutes a redundant set of partial concept descriptio ..."
Abstract

Cited by 77 (3 self)
 Add to MetaCart
This paper describes REGAL, a distributed genetic algorithmbased system, designed for learning First Order Logic concept descriptions from examples. The system is a hybrid between the Pittsburgh and the Michigan approaches, as the population constitutes a redundant set of partial concept descriptions, each evolved separately. In order to increase effectiveness, REGAL is specifically tailored to the concept learning task; hence, REGAL is taskdependent, but, on the other hand, domainindependent. The system proved to be particularly robust with respect to parameter setting across a variety of different application domains. REGAL is based on a selection operator, called Universal Suffrage operator, provably allowing the population to asymptotically converge, in average, to an equilibrium state, in which several species coexist. The system is presented both in a serial and in a parallel version, and a new distributed computational model is proposed and discussed. The system has been test...
Parallel Recombinative Simulated Annealing: A Genetic Algorithm
, 1995
"... This paper introduces and analyzes a parallel method of simulated annealing. Borrowing from genetic algorithms, an effective combination of simulated annealing and genetic algorithms, called parallel recombinative simulated annealing, is developed. This new algorithm strives to retain the desirable ..."
Abstract

Cited by 74 (3 self)
 Add to MetaCart
This paper introduces and analyzes a parallel method of simulated annealing. Borrowing from genetic algorithms, an effective combination of simulated annealing and genetic algorithms, called parallel recombinative simulated annealing, is developed. This new algorithm strives to retain the desirable asymptotic convergence properties of simulated annealing, while adding the populations approach and recombinative power of genetic algorithms. The algorithm iterates a population of solutions rather than a single solution, employing a binary recombination operator as well as a unary neighborhood operator. Proofs of global convergence are given for two variations of the algorithm. Convergence behavior is examined, and empirical distributions are compared to Boltzmann distributions. Parallel recombinative simulated annealing is amenable to straightforward implementation on SIMD, MIMD, or sharedmemory machines. The algorithm, implemented on the CM5, is run repeatedly on two deceptive problems...
Implicit Niching in a Learning Classifier System: Nature's Way
 EVOLUTIONARY COMPUTATION
, 1994
"... We approach the difficult task of analyzing the complex behavior of even the simplest learning classifier system (LCS) by isolating one crucial subfunction in the LCS learning algorithm: covering through niching. The LCS must maintain a population of diverse rules that together solve a problem (e.g. ..."
Abstract

Cited by 57 (9 self)
 Add to MetaCart
We approach the difficult task of analyzing the complex behavior of even the simplest learning classifier system (LCS) by isolating one crucial subfunction in the LCS learning algorithm: covering through niching. The LCS must maintain a population of diverse rules that together solve a problem (e.g., classify examples). To maintain a diverse population while applying the GA's selection operator, the LCS must incorporate some kind of niching mechanism. The natural way to accomplish niching in an LCS is to force competing rules to share resources (i.e., rewards). This implicit LCS fitness sharing is similar to the explicit fitness sharing used in many niched GAs. Indeed, the LCS implicit sharing algorithm can be mapped onto explicit fitness sharing with a onetoone correspondence between algorithm components. This mapping is important because several studies of explicit fitness sharing, and of niching in GAs generally, have produced key insights and analytical tools for understanding th...
Genetic Algorithms and the Variance of Fitness
, 1991
"... this paper, we consider one important source of stochastic variation, the variance of a schema's fitness or what we call collateral noise. Specifically, a method for calculating fitness variance from a function's Walsh transform is derived and applied to a number of problems in GA analysis. In the r ..."
Abstract

Cited by 48 (10 self)
 Add to MetaCart
this paper, we consider one important source of stochastic variation, the variance of a schema's fitness or what we call collateral noise. Specifically, a method for calculating fitness variance from a function's Walsh transform is derived and applied to a number of problems in GA analysis. In the remainder, Walsh functions and their application to the calculation of schema average fitness are reviewed; a formula for the calculation of schema fitness variance is derived using Walsh transforms. The variance computation is then applied to two important problems in genetic algorithm theory: population sizing and the calculation of rigorous probabilistic convergence bounds. Extending the technique to the analysis of nonuniform populations is also discussed. Re iew of als  c ema Anal sis