Abstract
Estimation-of-distribution algorithms (EDAs) are randomized search heuristics that create a probabilistic model of the solution space, which is updated iteratively, based on the quality of the solutions sampled according to the model. As previous works show, this iteration-based perspective can lead to erratic updates of the model, in particular, to bit-frequencies approaching a random boundary value. In order to overcome this problem, we propose a new EDA based on the classic compact genetic algorithm (cGA) that takes into account a longer history of samples and updates its model only with respect to information which it classifies as statistically significant. We prove that this significance-based cGA (sig-cGA) optimizes the commonly regarded benchmark functions OneMax (OM), LeadingOnes, and BinVal all in quasilinear time, a result shown for no other EDA or evolutionary algorithm so far. For the recently proposed stable compact genetic algorithm - an EDA that tries to prevent erratic model updates by imposing a bias to the uniformly distributed model - we prove that it optimizes OM only in a time exponential in its hypothetical population size. Similarly, we show that the convex search algorithm cannot optimize OM in polynomial time.
| Original language | English |
|---|---|
| Article number | 8917722 |
| Pages (from-to) | 1025-1034 |
| Number of pages | 10 |
| Journal | IEEE Transactions on Evolutionary Computation |
| Volume | 24 |
| Issue number | 6 |
| DOIs | |
| Publication status | Published - 1 Dec 2020 |
Keywords
- Estimation-of-distribution algorithm (EDA)
- run time analysis
- theory