We saw that there are several ways to do ES Gaussian Mutation: single global step, uncorrelated, and correlated. We can generalize this into just saying:
where:
- mean is the current search mean
- is the global step size
- is the covariance matrix that defines the shape and orientation of the search distribution
In 2 dimensions:
where
Recall that in classical correlated ES the chromosome was:
CMA-ES replaces explicit parameter mutation by learning the covariance matrix directly.
is typically learned through evolution paths; after sampling offspring, CMA-ES ranks them by fitness and keeps the best ones. Then:
- The mean is updated to move toward better samples
- The covariance matrix is updated to reinforce successful directions
- The global step size is increased when progress suggests larger moves, and decreased when refinement is needed.
CMA-ES is considered state-of-the-art; it’s much more principled than just correlated ES, because we are essentially still adaptively guessing in CMA-ES.