skip to main content
10.1145/3205455.3205487acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Analysis of information geometric optimization with isotropic gaussian distribution under finite samples

Published: 02 July 2018 Publication History

Abstract

In this article, we theoretically investigate the convergence properties of the information geometric optimization (IGO) algorithm given the family of isotropic Gaussian distributions on the sphere function. Differently from previous studies, where the exact natural gradient is taken, i.e., the infinite samples are assumed, we consider the case that the natural gradient is estimated from finite samples. We derive the rates of the expected decrease of the squared distance to the optimum and the variance parameter as functions of the learning rates, dimension, and sample size. From the rates of decrease deduces that the rates of decreases of the squared distance to the optimum and the variance parameter must agree for geometric convergence of the algorithm. In other words, the ratio between the squared distance to the optimum and the variance must be stable, which is observed empirically but is not derived in the previous theoretical studies. We further derive the condition on the learning rates that the rates of decreases agree and derive the stable value of the ratio. We confirm in simulation that the derived rates of decreases and the stable value of the ratio well approximate the behavior of the IGO algorithm.

References

[1]
Youhei Akimoto. 2012. Analysis of a Natural Gradient Algorithm on Monotonie Convex-Quadratic-Composite Functions. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO) 2012. 1293--1300.
[2]
Youhei Akimoto, Anne Auger, and Nikolaus Hansen. 2012. Convergence of the Continuous Time Trajectories of Isotropic Evolution Strategies on Monotonic C^2-composite Functions. In Parallel Problem Solving from Nature - PPSN XII, Part I (Lecture Notes in Computer Science), Vol. 7491. Springer, 42--51.
[3]
Youhei Akimoto, Anne Auger, and Nikolaus Hansen. 2017. Quality Gain Analysis of the Weighted Recombination Evolution Strategy on General Convex Quadratic Functions. In Proceedings of the 14th Conference on Foundations of Genetic Algorithms (FOGA) 2017. 111--126.
[4]
Youhei Akimoto, Anne Auger, and Nikolaus Hansen. 2017. Quality Gain Analysis of the Weighted Recombination Evolution Strategy on General Convex Quadratic Functions. arXiv:1604.00772 (2017). arXiv:1608.04813 https://arxiv.org/pdf/1608.04813.pdf
[5]
Shun-ichi Amari. 1998. Natural Gradient Works Efficiently in Learning. Neural Computation 10 (1998), 251--276.
[6]
Dirk V. Arnold. 2005. Optimal Weighted Recombination. In Proceedings of the 8th Conference on Foundations of Genetic Algorithms (FOGA) 2005. Springer, 215--237.
[7]
Anne Auger. 2005. Convergence results for the (1,λ)-SA-ES using the theory of φ-irreducible Markov chains. Theoretical Computer Science 334, 1-3 (2005), 35--69.
[8]
Shummet Baluja. 1994. Population-Based Incremental Learning: A Method for Integrating Genetic Search Based Function Optimization and Competitive Learning. Technical Report. Carnegie Mellon University Pittsburgh.
[9]
Hans Georg Beyer. 2014. Convergence Analysis of Evolutionary Algorithms That Are Based on the Paradigm of Information Geometry. Evolutionary Computation 22, 4 (2014), 679--709.
[10]
James F. Crow and Motoo Kimura. 1979. Efficiency of Truncation Selection. In Proceedings of the National Academy of Sciences of the United States of America, Vol. 76. 396--399.
[11]
Tobias Glasmachers. 2012. Convergence of the IGO-Flow of Isotropic Gaussian Distributions on Convex Quadratic Problems. In Parallel Problem Solving from Nature - PPSN XII, Part I (Lecture Notes in Computer Science), Vol. 7491. Springer, 1--10.
[12]
Nikolaus Hansen, Sibylle D. Müller, and Petros Koumoutsakos. 2003. Reducing the Time Complexity of the Derandomized Evolution Strategy with Covariance Matrix Adaptation (CMA-ES). Evolutionary Computation 11, 1 (2003), 1--18.
[13]
Silja Meyer-Nieberg and Hans-Georg Beyer. 2012. The Dynamical Systems Approach --- Progress Measures and Convergence Properties. Springer-Verlag Berlin Heidelberg, 741--814.
[14]
Yann Ollivier, Ludovic Arnold, Anne Auger, and Nikolaus Hansen. 2017. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. Journal of Machine Learning Research 18, 1 (2017), 564--628.

Cited By

View all
  • (2022)Anomaly Detection of Consumption in Hotel Units: A Case Study Comparing Isolation Forest and Variational Autoencoder AlgorithmsApplied Sciences10.3390/app1301031413:1(314)Online publication date: 27-Dec-2022
  • (2020)Finite-Sample Analysis of Information Geometric Optimization With Isotropic Gaussian Distribution on Convex Quadratic FunctionsIEEE Transactions on Evolutionary Computation10.1109/TEVC.2019.291770924:6(1035-1049)Online publication date: 1-Dec-2020
  • (2019)Large-scale noise-resilient evolution-strategiesProceedings of the Genetic and Evolutionary Computation Conference10.1145/3321707.3321724(682-690)Online publication date: 13-Jul-2019

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
GECCO '18: Proceedings of the Genetic and Evolutionary Computation Conference
July 2018
1578 pages
ISBN:9781450356183
DOI:10.1145/3205455
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 July 2018

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. convergence property
  2. finite samples
  3. information geometric optimization
  4. natural gradient
  5. theory

Qualifiers

  • Research-article

Conference

GECCO '18
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 02 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Anomaly Detection of Consumption in Hotel Units: A Case Study Comparing Isolation Forest and Variational Autoencoder AlgorithmsApplied Sciences10.3390/app1301031413:1(314)Online publication date: 27-Dec-2022
  • (2020)Finite-Sample Analysis of Information Geometric Optimization With Isotropic Gaussian Distribution on Convex Quadratic FunctionsIEEE Transactions on Evolutionary Computation10.1109/TEVC.2019.291770924:6(1035-1049)Online publication date: 1-Dec-2020
  • (2019)Large-scale noise-resilient evolution-strategiesProceedings of the Genetic and Evolutionary Computation Conference10.1145/3321707.3321724(682-690)Online publication date: 13-Jul-2019

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media