Diagonal Acceleration for Covariance Matrix Adaptation Evolution Strategies

05/14/2019
by   Youhei Akimoto, et al.
0

We introduce an acceleration for covariance matrix adaptation evolution strategies (CMA-ES) by means of adaptive diagonal decoding (dd-CMA). This diagonal acceleration endows the default CMA-ES with the advantages of separable CMA-ES without inheriting its drawbacks. Technically, we introduce a diagonal matrix D that expresses coordinate-wise variances of the sampling distribution in DCD form. The diagonal matrix can learn a rescaling of the problem in the coordinates within linear number of function evaluations. Diagonal decoding can also exploit separability of the problem, but, crucially, does not compromise the performance on non-separable problems. The latter is accomplished by modulating the learning rate for the diagonal matrix based on the condition number of the underlying correlation matrix. dd-CMA-ES not only combines the advantages of default and separable CMA-ES, but may achieve overadditive speedup: it improves the performance, and even the scaling, of the better of default and separable CMA-ES on classes of non-separable test functions that reflect, arguably, a landscape feature commonly observed in practice. The paper makes two further secondary contributions: we introduce two different approaches to guarantee positive definiteness of the covariance matrix with active CMA, which is valuable in particular with large population size; we revise the default parameter setting in CMA-ES, proposing accelerated settings in particular for large dimension. All our contributions can be viewed as independent improvements of CMA-ES, yet they are also complementary and can be seamlessly combined. In numerical experiments with dd-CMA-ES up to dimension 5120, we observe remarkable improvements over the original covariance matrix adaptation on functions with coordinate-wise ill-conditioning. The improvement is observed also for large population sizes up to about dimension squared.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/25/2022

Core Shrinkage Covariance Estimation for Matrix-variate Data

A separable covariance model for a random matrix provides a parsimonious...
research
08/17/2018

Inconsistency of diagonal scaling under high-dimensional limit: a replica approach

In this note, we claim that diagonal scaling of a sample covariance matr...
research
11/01/2022

Fluctuations of the diagonal entries of a large sample precision matrix

For a given p× n data matrix X_n with i.i.d. centered entries and a popu...
research
04/07/2023

CMA-ES with Learning Rate Adaptation: Can CMA-ES with Default Population Size Solve Multimodal and Noisy Problems?

The covariance matrix adaptation evolution strategy (CMA-ES) is one of t...
research
06/10/2014

Maximum Likelihood-based Online Adaptation of Hyper-parameters in CMA-ES

The Covariance Matrix Adaptation Evolution Strategy (CMA-ES) is widely a...
research
06/10/2011

A Linear Time Natural Evolution Strategy for Non-Separable Functions

We present a novel Natural Evolution Strategy (NES) variant, the Rank-On...
research
06/23/2016

On the Theoretical Capacity of Evolution Strategies to Statistically Learn the Landscape Hessian

We study the theoretical capacity to statistically learn local landscape...

Please sign up or login with your details

Forgot password? Click here to reset