We saw that there are several ways to do ES Gaussian Mutation: single global step, uncorrelated, and correlated. We can generalize this into just saying:

where:

  • mean is the current search mean
  • is the global step size
  • is the covariance matrix that defines the shape and orientation of the search distribution

In 2 dimensions:

where

Recall that in classical correlated ES the chromosome was:

CMA-ES replaces explicit parameter mutation by learning the covariance matrix directly.

is typically learned through evolution paths; after sampling offspring, CMA-ES ranks them by fitness and keeps the best ones. Then:

  • The mean is updated to move toward better samples
  • The covariance matrix is updated to reinforce successful directions
  • The global step size is increased when progress suggests larger moves, and decreased when refinement is needed.

CMA-ES is considered state-of-the-art; it’s much more principled than just correlated ES, because we are essentially still adaptively guessing in CMA-ES.