Scaling of Probability-Based Optimization Algorithms
Abstract
Population-based Incremental Learning is shown require very sen(cid:173) sitive scaling of its learning rate. The learning rate must scale with the system size in a problem-dependent way. This is shown in two problems: the needle-in-a haystack, in which the learning rate must vanish exponentially in the system size, and in a smooth function in which the learning rate must vanish like the square root of the system size. Two methods are proposed for removing this sensitiv(cid:173) ity. A learning dynamics which obeys detailed balance is shown to give consistent performance over the entire range of learning rates. An analog of mutation is shown to require a learning rate which scales as the inverse system size, but is problem independent.
Cite
Text
Shapiro. "Scaling of Probability-Based Optimization Algorithms." Neural Information Processing Systems, 2002.Markdown
[Shapiro. "Scaling of Probability-Based Optimization Algorithms." Neural Information Processing Systems, 2002.](https://mlanthology.org/neurips/2002/shapiro2002neurips-scaling/)BibTeX
@inproceedings{shapiro2002neurips-scaling,
title = {{Scaling of Probability-Based Optimization Algorithms}},
author = {Shapiro, J. L.},
booktitle = {Neural Information Processing Systems},
year = {2002},
pages = {399-406},
url = {https://mlanthology.org/neurips/2002/shapiro2002neurips-scaling/}
}