Abstract
This paper presents Natural Evolution Strategies (NES), a recent family of black-box optimization algorithms that use the natural gradient to update a parameterized search distribution in the direction of higher expected fitness. We introduce a collection of techniques that address issues of convergence, robustness, sample complexity, computational complexity and sensitivity to hyperparameters. This paper explores a number of implementations of the NES family, such as general-purpose multi-variate normal distributions and separable distributions tailored towards search in high dimensional spaces. Experimental results show best published performance on various standard benchmarks, as well as competitive performance on others. © 2014 Daan Wierstra, Tom Schaul, Tobias Glasmachers, Yi Sun, Jan Peters and Jürgen Schmidhuber.
Original language | English (US) |
---|---|
Pages (from-to) | 949-980 |
Number of pages | 32 |
Journal | Journal of Machine Learning Research |
Volume | 15 |
State | Published - Jan 1 2014 |
Externally published | Yes |
ASJC Scopus subject areas
- Artificial Intelligence
- Software
- Statistics and Probability
- Control and Systems Engineering