|
|
Prof.
Dr. rer. nat. habil. Hans-Georg Beyer
|
Recent Talks
1. Evolution Strategies are Not Gradient Followers
In order to explain the working principles of Evolution Strategies (ESs)
in real-valued search spaces, sometimes the picture of a (stochastic)
gradient approximating strategy is invoked. There are publications in
the field of machine learning and evolutionary algorithms where this
misleading picture is promoted. Therefore, I gave a talk at
Dagstuhl,
Seminar 19431 (Oct. 20 - 25, 2019),
showing that this picture is not correct: ESs are much more explorative than
gradient strategies, thus they have a certain chance of not being trapped in the
next local attractor. The slides of that talk can
be obtained here.
BTW, even the consideration of ESs as Monte-Carlo approximators of the
so-called natural gradient does not hold for standard ESs such as the
Covariance Matrix Adaptation ES. A discussion of that topic can be found in my paper
Convergence Analysis of Evolutionary Algorithms
That are Based on the Paradigm of Information Geometry. While the main
part of that paper is rather technical, the Introduction and the Conclusions
should be easy to follow.
2. Design Principles for Matrix Adaptation Evolution Strategies
In the paper Simplify Your
Covariance Matrix Adaptation Evolution Strategy we have shown that one can
simplify this well-performing ES by removing the covariance matrix totally
from the CMA-ES without performance degradation. As a result one obtains simpler
Evolution Strategies that allow for further algorithm engineering addressing
high-dimensional search spaces and constrained optimization problems.
Here are tutorial slides
discussing these topics.
Matlab/Octave code of the basic algorithms can be found at
Downloads
This talk held in Malaga, Spain, at the GECCO 2025
AABOH Workshop
on July 14th, 2025, takes a critical look at the
well-established
COCO BBOB
benchmark suite.
After a short recap of the ideas behind ECDF plots (empirical cumulative
distribution function) a couple of weaknesses of the COCO BBOB design are
discussed:
- The function value targets between 10-6 and 10-8
are far too precise for real-world applications. These only reflect the local convergence behavior of the algorithms evaluated.
- Aggregating function value targets of different test functions in a single ECDF plot is scientific nonsense.
- The transformations used to generate different instantiations of the test functions favor CMA-ES like algorithms.
- The 24 test functions used need a renovation in order to reflect the developments of the last 15 years.
Regarding the last item of this bullet list, I've proposed a couple of
simple, scale- and tunable, but challenging test functions that appear hard for
CMA-ES like algorithms (and others). These are HappyCat-, HGBat-, ThreeButton-,
and the Audi-function. The peculiarity of these functions is that (except for
Audi) they have only one local minimum (being also the global one). In 2D one
can easily locate the global minimizer visually. Yet, it is
difficult for the algorithms to get close to the global minimizer.
For example, in the case of ThreeButton it takes the CMA-ES very long to
get to the minimizer even in the case of a rather weak 2D version of
this function. This can be seen in this
mp4-video
where the offspring population (pink dots) is displayed in the search space
along with the
contour lines of the ThreeButton function. Unlike most of the functions in the
BBOB test suite where the CMA-ES has to learn the covariance matrix ones, in
the case of ThreeButton it has to permanently learn and re-learn
this matrix (note, in order to shorten the video, some generation intervals of
the evolution process were skipped).
In the slides of this talk additional videos have been linked to show the
evolution of the population in the HappyCat and HGBat landscape for 2D and
10D instances. For the
higher-dimensional case a visualization trick has been invented to display the
exploratory behavior of the offspring population. To this end, those two
dimensions of the parent centroid that are most far apart from the minimizer
are determined. These two dimensions (that can change from generation to
generation) are used to display the offspring individuals. As an example, see the
(simple)
10D HGBat case with β=1.
When thinking about scaleable test functions in a renovated BBOB test suite also
examples from physics come into mind. Some of them are discussed in this talk:
Thomson's and Tammes' problem as well as the Lennard-Jones energy minimization.
All these problems are well-suited for large-scale optimization benchmarking.
last change: 22.02.2026
|