skip to main content
10.1145/2576768.2598258acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Comparison-based natural gradient optimization in high dimension

Published: 12 July 2014 Publication History

Abstract

We propose a novel natural gradient based stochastic search algorithm, VD-CMA, for the optimization of high dimensional numerical functions. The algorithm is comparison-based and hence invariant to monotonic transformations of the objective function. It adapts a multivariate normal distribution with a restricted covariance matrix with twice the dimension as degrees of freedom, representing an arbitrarily oriented long axis and additional axis-parallel scaling. We derive the different components of the algorithm and show linear internal time and space complexity. We find empirically that the algorithm adapts its covariance matrix to the inverse Hessian on convex-quadratic functions with an Hessian with one short axis and different scaling on the diagonal. We then evaluate VD-CMA on test functions and compare it to different methods. On functions covered by the internal model of VD-CMA and on the Rosenbrock function, VD-CMA outperforms CMA-ES (having quadratic internal time and space complexity) not only in internal complexity but also in number of function calls with increasing dimension.

References

[1]
Y. Akimoto, Y. Nagata, I. Ono, and S. Kobayashi. Bidirectional Relation between CMA Evolution Strategies and Natural Evolution Strategies. In Parallel Problem Solving from Nature -- PPSN XI, pages 154--163, 2010.
[2]
Y. Akimoto, Y. Nagata, I. Ono, and S. Kobayashi. Theoretical Foundation for CMA-ES from Information Geometry Perspective. Algorithmica, 64:698--716, 2012.
[3]
S. Amari. Natural Gradient Works Efficiently in Learning. Neural Computation, 10(2):251--276, 1998.
[4]
S. Amari, H. Park, and K. Fukumizu. Adaptive method of realizing natural gradient learning for multilayer perceptrons. Neural Computation, 12:1399--1409, 2000.
[5]
Y. Ollivier, L. Arnold, A. Auger, and N. Hansen. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. arXiv:1106.3708, 2011.
[6]
P.-T. D. Boer, D. P. Kroese, S. Mannor, and R. Y. Rubinstein. A Tutorial on the Cross-Entropy Method. Annals of Operations Research, (134):19--67, 2005.
[7]
T. Glasmachers, T. Schaul, Y. Sun, D. Wierstra, and J. Schmidhuber. Exponential Natural Evolution Strategies. In Proceedings of Genetic and Evolutionary Computation Conference, pages 393--400, 2010.
[8]
N. Hansen and A. Auger. Principled Design of Continuous Stochastic Search: From Theory to Practice. In Y. Borenstein and A. Moraglio, editors, Theory and Principled Methods for the Design of Metaheuristics. Springer, 2013.
[9]
N. Hansen, S. D. Muller, and P. Koumoutsakos. Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES). Evolutionary Computation, 11(1):1--18, 2003.
[10]
N. Hansen and A. Ostermeier. Completely derandomized self-adaptation in evolution strategies. Evolutionary Computation, 9(2):159--195, 2001.
[11]
N. Hansen, A. Ostermeier, and A. Gawelczyk. On the Adaptation of Arbitrary Normal Mutation Distributions in Evolution Strategies: The Generating set Adaptation. In Proceedings of the Sixth International Conference on Genetic Algorithms, pages 57--64, 1995.
[12]
D. A. Harville. Matrix Algebra from a Statistician's Perspective. Springer-Verlag, 2008.
[13]
A. Honkela, M. Tornio, T. Raiko, and J. Karhunen. Natural conjugate gradient in variational inference. In Neural Information Processing, 14th International Conference, ICONIP 2007, pages 305--314, 2008.
[14]
L. Malagò, M. Matteucci, and G. Pistone. Towards the geometry of estimation of distribution algorithms based on the exponential family. In Foundations of Genetic Algorithms, pages 230--242. 2011.
[15]
A. Miyamae, Y. Nagata, I. Ono, and S. Kobayashi. Natural Policy Gradient Methods with Parameter-based Exploration for Control Tasks. In Advances in Neural Information Processing Systems 23, pages 1660--1668, 2010.
[16]
J. Peters and S. Schaal. Natural actor-critic. Neurocomputing, 71(7--9):1180--1190, 2008.
[17]
J. Poland and A. Zell. Main vector adaptation: A CMA variant with linear time and space complexity. In Proceedings of the Genetic and Evolutionary Computation Conference, pages 1050--1055, 2001.
[18]
R. Ros and N. Hansen. A simple modification in CMA-ES achieving linear time and space complexity. Parallel Problem Solving from Nature--PPSN X, pages 296--305, 2008.
[19]
Y. Sun, F. Gomez, T. Schaul, and J. Schmidhuber. A Linear Time Natural Evolution Strategy for Non-Separable Functions. GECCO'13 Companion, pages 61--62, 2013.
[20]
T. Suttorp, N. Hansen, and C. Igel. Efficient covariance matrix update for variable metric evolution strategies. Machine Learning, 75(2):167--197, 2009.
[21]
D. Wierstra, T. Schaul, J. Peters, and J. Schmidhuber. Natural evolution strategies. In IEEE Congress on Evolutionary Computation, pages 3381--3387, 2008.

Cited By

View all
  • (2025)Exploring high-dimensional optimization by sparse and low-rank evolution strategySwarm and Evolutionary Computation10.1016/j.swevo.2024.10182892(101828)Online publication date: Feb-2025
  • (2023)A Dynamic Partial Update for Covariance Matrix AdaptationProceedings of the Companion Conference on Genetic and Evolutionary Computation10.1145/3583133.3596385(2394-2397)Online publication date: 15-Jul-2023
  • (2022)Benchmarking an algorithm for expensive high-dimensional objectives on the bbob and bbob-largescale testbedsProceedings of the Genetic and Evolutionary Computation Conference Companion10.1145/3520304.3534006(1725-1733)Online publication date: 9-Jul-2022
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
GECCO '14: Proceedings of the 2014 Annual Conference on Genetic and Evolutionary Computation
July 2014
1478 pages
ISBN:9781450326629
DOI:10.1145/2576768
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 July 2014

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. covariance matrix adaptation
  2. hessian matrix
  3. information geometry
  4. natural gradient
  5. theory

Qualifiers

  • Research-article

Funding Sources

Conference

GECCO '14
Sponsor:
GECCO '14: Genetic and Evolutionary Computation Conference
July 12 - 16, 2014
BC, Vancouver, Canada

Acceptance Rates

GECCO '14 Paper Acceptance Rate 180 of 544 submissions, 33%;
Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)30
  • Downloads (Last 6 weeks)1
Reflects downloads up to 02 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Exploring high-dimensional optimization by sparse and low-rank evolution strategySwarm and Evolutionary Computation10.1016/j.swevo.2024.10182892(101828)Online publication date: Feb-2025
  • (2023)A Dynamic Partial Update for Covariance Matrix AdaptationProceedings of the Companion Conference on Genetic and Evolutionary Computation10.1145/3583133.3596385(2394-2397)Online publication date: 15-Jul-2023
  • (2022)Benchmarking an algorithm for expensive high-dimensional objectives on the bbob and bbob-largescale testbedsProceedings of the Genetic and Evolutionary Computation Conference Companion10.1145/3520304.3534006(1725-1733)Online publication date: 9-Jul-2022
  • (2022)DiBBProceedings of the Genetic and Evolutionary Computation Conference10.1145/3512290.3528764(341-349)Online publication date: 8-Jul-2022
  • (2022)Fast Moving Natural Evolution Strategy for High-Dimensional Problems2022 IEEE Congress on Evolutionary Computation (CEC)10.1109/CEC55065.2022.9870206(1-8)Online publication date: 18-Jul-2022
  • (2022)Analysis of Surrogate-Assisted Information-Geometric Optimization AlgorithmsAlgorithmica10.1007/s00453-022-01087-886:1(33-63)Online publication date: 22-Dec-2022
  • (2022)Towards a Principled Learning Rate Adaptation for Natural Evolution StrategiesApplications of Evolutionary Computation10.1007/978-3-031-02462-7_45(721-737)Online publication date: 15-Apr-2022
  • (2021)Convergence rate of the (1+1)-evolution strategy with success-based step-size adaptation on convex quadratic functionsProceedings of the Genetic and Evolutionary Computation Conference10.1145/3449639.3459289(1169-1177)Online publication date: 26-Jun-2021
  • (2021)MMES: Mixture Model-Based Evolution Strategy for Large-Scale OptimizationIEEE Transactions on Evolutionary Computation10.1109/TEVC.2020.303476925:2(320-333)Online publication date: Apr-2021
  • (2021)Large-Scale Evolution Strategy Based on Search Direction AdaptationIEEE Transactions on Cybernetics10.1109/TCYB.2019.292856351:3(1651-1665)Online publication date: Mar-2021
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media