Results 1  10
of
70
Hierarchical Bayesian Optimization Algorithm = Bayesian Optimization Algorithm + Niching + Local Structures
, 2001
"... The paper describes the hierarchical Bayesian optimization algorithm which combines the Bayesian optimization algorithm, local structures in Bayesian networks, and a powerful niching technique. The proposed algorithm is able to solve hierarchical traps and other difficult problems very efficiently. ..."
Abstract

Cited by 255 (63 self)
 Add to MetaCart
The paper describes the hierarchical Bayesian optimization algorithm which combines the Bayesian optimization algorithm, local structures in Bayesian networks, and a powerful niching technique. The proposed algorithm is able to solve hierarchical traps and other difficult problems very efficiently.
Hierarchical BOA Solves Ising Spin Glasses and MAXSAT
 In Proc. of the Genetic and Evolutionary Computation Conference (GECCO 2003), number 2724 in LNCS
, 2003
"... Theoretical and empirical evidence exists that the hierarchical Bayesian optimization algorithm (hBOA) can solve challenging hierarchical problems and anything easier. This paper applies hBOA to two important classes of realworld problems: Ising spinglass systems and maximum satis ability (MAX ..."
Abstract

Cited by 46 (17 self)
 Add to MetaCart
Theoretical and empirical evidence exists that the hierarchical Bayesian optimization algorithm (hBOA) can solve challenging hierarchical problems and anything easier. This paper applies hBOA to two important classes of realworld problems: Ising spinglass systems and maximum satis ability (MAXSAT). The paper shows how easy it is to apply hBOA to realworld optimization problems. The results indicate that hBOA is capable of solving enormously dicult problems that cannot be solved by other optimizers and still provide competitive or better performance than problemspeci c approaches on other problems. The results thus con rm that hBOA is a practical, robust, and scalable technique for solving challenging realworld problems.
Probabilistic Model Building and Competent Genetic Programming
 GENETIC PROGRAMMING THEORY AND PRACTISE, CHAPTER 13
, 2003
"... This paper describes a probabilistic model building genetic programming (PMBGP) developed based on the extended compact genetic algorithm (eCGA). Unlike traditional genetic programming, which use fixed recombination operators, the proposed PMBGA adapts linkages. The proposed algorithms... ..."
Abstract

Cited by 39 (10 self)
 Add to MetaCart
This paper describes a probabilistic model building genetic programming (PMBGP) developed based on the extended compact genetic algorithm (eCGA). Unlike traditional genetic programming, which use fixed recombination operators, the proposed PMBGA adapts linkages. The proposed algorithms...
Fitness inheritance in the Bayesian optimization algorithm
, 2004
"... This paper describes how fitness inheritance can be used to estimate fitness for a proportion of newly sampled candidate solutions in the Bayesian optimization algorithm (BOA). The goal of estimating fitness for some candidate solutions is to reduce the number of fitness evaluations for problems whe ..."
Abstract

Cited by 31 (22 self)
 Add to MetaCart
This paper describes how fitness inheritance can be used to estimate fitness for a proportion of newly sampled candidate solutions in the Bayesian optimization algorithm (BOA). The goal of estimating fitness for some candidate solutions is to reduce the number of fitness evaluations for problems where fitness evaluation is expensive. Bayesian networks used in BOA to model promising solutions and generate the new ones are extended to allow not only for modeling and sampling candidate solutions, but also for estimating their fitness. The results indicate that fitness inheritance is a promising concept in BOA, because populationsizing requirements for building appropriate models of promising solutions lead to good fitness estimates even if only a small proportion of candidate solutions is evaluated using the actual fitness function. This can lead to a reduction of the number of actual fitness evaluations by a factor of 30 or more.
Designing competent mutation operators via probabilistic model building of neighborhoods
 In Deb, K., & et al. (Eds.), Proceedings of the Genetic and Evolutionary Computation Conference (GECCO2004), Part II, LNCS 3103
, 2004
"... This paper presents a competent selectomutative genetic algorithm (GA), that adapts linkage and solves hard problems quickly, reliably, and accurately. A probabilistic model building process is used to automatically identify key building blocks (BBs) of the search problem. The mutation operator uses ..."
Abstract

Cited by 31 (21 self)
 Add to MetaCart
This paper presents a competent selectomutative genetic algorithm (GA), that adapts linkage and solves hard problems quickly, reliably, and accurately. A probabilistic model building process is used to automatically identify key building blocks (BBs) of the search problem. The mutation operator uses the probabilistic model of linkage groups to find the best among competing building blocks. The competent selectomutative GA successfully solves additively separable problems of bounded difficulty, requiring only subquadratic number of function evaluations. The results show that for additively separable problems the probabilistic model building BBwise mutation scales as O(2 k m 1.5), and requires O ( √ k log m) less function evaluations than its selectorecombinative counterpart, confirming theoretical results reported elsewhere (Sastry & Goldberg, 2004). 1
Using Genetic Algorithms for
 Concept Learning, Machine Learning
, 1993
"... Abstract. Current Genetic Algorithms can efficiently address orderk separable problems, in which the order of the linkage is restricted to a low value k. Outside this class, there exist hierarchical problems that cannot be addressed by current genetic algorithms, yet can be addressed efficiently in ..."
Abstract

Cited by 25 (2 self)
 Add to MetaCart
Abstract. Current Genetic Algorithms can efficiently address orderk separable problems, in which the order of the linkage is restricted to a low value k. Outside this class, there exist hierarchical problems that cannot be addressed by current genetic algorithms, yet can be addressed efficiently in principle by exploiting hierarchy. We delineate the class of hierarchical problems, and describe a framework for Hierarchical Genetic Algorithms. Based on this outline for algorithms, we investigate under what conditions hierarchical problems may be solved efficiently. Sufficient conditions are provided under which hierarchical problems can be addressed in polynomial time. The analysis points to the importance of efficient sampling techniques that assess the quality of module settings. 1
Parallel estimation of distribution algorithms
, 2002
"... The thesis deals with the new evolutionary paradigm based on the concept of Estimation of Distribution Algorithms (EDAs) that use probabilistic model of promising solutions found so far to obtain new candidate solutions of optimized problem. There are six primary goals of this thesis: 1. Suggestion ..."
Abstract

Cited by 22 (3 self)
 Add to MetaCart
The thesis deals with the new evolutionary paradigm based on the concept of Estimation of Distribution Algorithms (EDAs) that use probabilistic model of promising solutions found so far to obtain new candidate solutions of optimized problem. There are six primary goals of this thesis: 1. Suggestion of a new formal description of EDA algorithm. This high level concept can be used to compare the generality of various probabilistic models by comparing the properties of underlying mappings. Also, some convergence issues are discussed and theoretical ways for further improvements are proposed. 2. Development of new probabilistic model and methods capable of dealing with continuous parameters. The resulting Mixed Bayesian Optimization Algorithm (MBOA) uses a set of decision trees to express the probability model. Its main advantage against the mostly used IDEA and EGNA approach is its backward compatibility with discrete domains, so it is uniquely capable of learning linkage between mixed continuousdiscrete genes. MBOA handles the discretization of continuous parameters as an integral part of the learning process, which outperforms the histogrambased
Representation development from ParetoCoevolution
 In Genetic and Evolutionary Computation Conference (GECCO
, 2003
"... Abstract. Genetic algorithms generally use a fixed problem representation that maps variables of the search space to variables of the problem, and operators of variation that are fixed over time. This limits their scalability on nonseparable problems. To address this issue, methods have been propos ..."
Abstract

Cited by 21 (3 self)
 Add to MetaCart
Abstract. Genetic algorithms generally use a fixed problem representation that maps variables of the search space to variables of the problem, and operators of variation that are fixed over time. This limits their scalability on nonseparable problems. To address this issue, methods have been proposed that coevolve explicitly represented modules. An open question is how modules in such coevolutionary setups should be evaluated. Recently, Paretocoevolution has provided a theoretical basis for evaluation in coevolution. We define a notion of functional modularity, and objectives for module evaluation based on ParetoCoevolution. It is shown that optimization of these objectives maximizes functional modularity. The resulting evaluation method is developed into an algorithm for variable length, open ended development of representations called DevRep. DevRep successfully identifies large partial solutions and greatly outperforms fixed length and variable length genetic algorithms on several test problems, including the 1024bit HierarchicalXOR problem.
Efficiency enhancement of genetic algorithms via buildingblockwise fitness estimation
 PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON EVOLUTIONARY COMPUTATION
, 2004
"... This paper studies fitness inheritance as an efficiency enhancement technique for a class of competent genetic algorithms called estimation distribution algorithms. Probabilistic models of important subsolutions are developed to estimate the fitness of a proportion of individuals in the population, ..."
Abstract

Cited by 21 (17 self)
 Add to MetaCart
This paper studies fitness inheritance as an efficiency enhancement technique for a class of competent genetic algorithms called estimation distribution algorithms. Probabilistic models of important subsolutions are developed to estimate the fitness of a proportion of individuals in the population, thereby avoiding computationally expensive function evaluations. The effect of fitness inheritance on the convergence time and population sizing are modeled and the speedup obtained through inheritance is predicted. The results show that a fitnessinheritance mechanism which utilizes information on buildingblock fitnesses provides significant efficiency enhancement. For additively separable problems, fitness inheritance reduces the number of function evaluations to about half and yields a speedup of about 1.75–2.25.
Analyzing probabilistic models in hierarchical boa on traps and spin glasses
 Genetic and Evolutionary Computation Conference (GECCO2007), I
, 2007
"... The hierarchical Bayesian optimization algorithm (hBOA) can solve nearly decomposable and hierarchical problems of bounded difficulty in a robust and scalable manner by building and sampling probabilistic models of promising solutions. This paper analyzes probabilistic models in hBOA on two common t ..."
Abstract

Cited by 17 (15 self)
 Add to MetaCart
The hierarchical Bayesian optimization algorithm (hBOA) can solve nearly decomposable and hierarchical problems of bounded difficulty in a robust and scalable manner by building and sampling probabilistic models of promising solutions. This paper analyzes probabilistic models in hBOA on two common test problems: concatenated traps and 2D Ising spin glasses with periodic boundary conditions. We argue that although Bayesian networks with local structures can encode complex probability distributions, analyzing these models in hBOA is relatively straightforward and the results of such analyses may provide practitioners with useful information about their problems. The results show that the probabilistic models in hBOA closely correspond to the structure of the underlying problem, the models do not change significantly in subsequent iterations of BOA, and creating adequate probabilistic models by hand is not straightforward even with complete knowledge of the optimization problem. Categories and Subject Descriptors