@inproceedings{teytaud:inria-00451416,
hal_id = {inria-00451416},
url = {http://hal.inria.fr/inria-00451416},
title = {{Bias and variance in continuous EDA}},
author = {Teytaud, Fabien and Teytaud, Olivier},
abstract = {{Estimation of Distribution Algorithms are based on statistical estimates. We show that when combining classical tools from statistics, namely bias/variance decomposition, reweighting and quasi-randomization, we can strongly improve the convergence rate. All modifications are easy, compliant with most algorithms, and experimentally very efficient in particular in the parallel case (large offsprings).}},
language = {Anglais},
affiliation = {TAO - INRIA Futurs , Laboratoire de Recherche en Informatique - LRI , TAO - INRIA Saclay - Ile de France},
booktitle = {{EA 09}},
address = {Strasbourg, France},
audience = {internationale },
year = {2009},
month = May,
pdf = {http://hal.inria.fr/inria-00451416/PDF/decsigma.pdf},
}
1. Why one must
use reweighting
F. Teytaud, O. Teytaud
Montréal, 2009
Tao, Inria Saclay Ile-De-France, LRI (Université Paris Sud, France),
UMR CNRS 8623, I&A team, Digiteo, Pascal Network of Excellence.
2. Outline
Idea of averaging in evolutionary
algorithms
This idea introduces a bias
How to remove this bias
The results
Conclusions
Teytaud and Teytaud Gecco 09 is great 2
3. Idea in ES
Average of selected points
= good approximation of
optimum
Teytaud and Teytaud Gecco 09 is great 3
4. Idea in ES
Let's assume this is true (for the
moment)...
nonetheless, there's a bias.
Teytaud and Teytaud Gecco 09 is great 4
5. EMNA (P. Larranaga and J.-A. Lozano, 2001)
While (not finished)
- generate population
- select best individuals
- estimate mean / variance
(and possibly covariance)
Teytaud and Teytaud Gecco 09 is great 5
6. EMNA (P. Larranaga and J.-A. Lozano, 2001)
Teytaud and Teytaud Gecco 09 is great 6
7. EMNA (P. Larranaga and J.-A. Lozano, 2001)
While (not finished)
- generate population
- select best individuals
- estimate mean / variance (and possibly covariance)
Highly parallel (more than most ES; T. et al, EvoStar 2001)
Very simple
Can handle covariance matrix easily
Teytaud and Teytaud Gecco 09 is great 7
8. Please wake up during 3 slides :-)
Idea of averaging in evolutionary algorithms
This idea introduces a bias
How to remove this bias
The results
Conclusions
Teytaud and Teytaud Gecco 09 is great 8
9. Bias due to bad (Gaussian) distribution
High
density
Teytaud and Teytaud Gecco 09 is great 9
10. Bias due to bad distribution
Teytaud and Teytaud Gecco 09 is great 10
11. Bias due to bad distribution
Teytaud and Teytaud Gecco 09 is great 11
12. Bias due to bad distribution
AVERAGE
(biased by the
distribution)
Teytaud and Teytaud Gecco 09 is great 12
13. Corrected by weighted average
AVERAGE
(the one we really
want !)
Teytaud and Teytaud Gecco 09 is great 13
14. Outline
Idea of averaging in evolutionary algorithms
This idea introduces a bias
How to remove this bias
The results
Conclusions
Teytaud and Teytaud Gecco 09 is great 14
15. American Election of 1936 (fun)
Literary digest: pop size = 2 000 000
==> predicts Landon
Gallup: pop size = 50 000
==> predicts Roosevelt (and was proved right)
Teytaud and Teytaud Gecco 09 is great 15
16. American Election of 1936
Literary digest: pop size = 2 300 000
==> predicts Landon
Gallup: pop size = 50 000
==> predicts Roosevelt (and was proved right)
The Literary digest failed because of a biased sampling.
(much more affluent people and much more republicans
among Literary Digest readers)
Correction: Weight of individual = real density / biased density.
Teytaud and Teytaud Gecco 09 is great 16
17. REMNA (reweighted EMNA)
Inverse Gaussian density
Gaussian density (=weight for removing the bias!)
Teytaud and Teytaud Gecco 09 is great 17
18. REMNA (reweighted EMNA)
Very simple modification:
- compute weight ( individual ) = 1 / density
- compute mean, variance, covariance
with these weights
==> not only for Gaussians
==> ok for all surrogate models / EDA
==> just an application of standard statistics
Teytaud and Teytaud Gecco 09 is great 18
20. Outline
Idea of averaging in evolutionary algorithms
This idea introduces a bias
How to remove this bias
The results: less premature convergence
Conclusions
Teytaud and Teytaud Gecco 09 is great 20
21. The results
We do not prove that
“there's no more premature convergence.”
We just show that, for a fixed generation,
“ IF the center of the level set is the optimum,
THEN
the asymptotic value of the estimated optimum
= the optimum.”
==> is the condition really necessary ?
Teytaud and Teytaud Gecco 09 is great 21
22. Yes: center = optimum and Yes: situation better
Yes and yes
Teytaud and Teytaud Gecco 09 is great 22
23. No: center ≠optimum but Yes: situation better
(assumption not really
necessary)
No... but yes
Teytaud and Teytaud Gecco 09 is great 23
25. Conclusions
Idea of averaging in evolutionary algorithms
This idea introduces a bias
How to remove this bias
The results
Conclusions
Teytaud and Teytaud Gecco 09 is great 25
26. Conclusions
Reduces the risk of premature convergence
No proof on the complete algorithm (just
step-wise consistency)
Empirically quite good for EMNA (should be
tested on other EDA / surrogate)
Simple, sound, widely applicable
Bias of step-size adaptation not yet
analyzed (==> seemingly works quite well!)
Teytaud and Teytaud Gecco 09 is great 26
27. Related work
Papers from D.V. Arnold et al. around
reweighting for improved convergence
rate (sphere, ridge) of ES
(to be combined ?)
Work from CMA-people around weights for
improved conv. rate in CMA-ES
Thanks! Questions ?
Teytaud and Teytaud Gecco 09 is great 27
Notas del editor
I am Frederic Lemoine, PhD student at the University Paris Sud. I will present you my work on GenoQuery, a new querying module adapted to a functional genomics warehouse