Section: Research Program
Developing Novel Theoretical Frameworks for Analyzing and Designing Adaptive Stochastic Algorithms
The lines of research of the RandOpt team are organized along four axis namely developing novel theoretical framework, developing novel algorithms, setting novel standards in scientific experimentation and benchmarking and applications.
Stochastic blackbox algorithms typically optimize nonconvex, nonsmooth functions. This is possible because the algorithms rely on weak mathematical properties of the underlying functions: not only derivatives (gradients) are not exploited, but often the methods are socalled comparisonbased which means that the algorithm will only rely on the ranking of the candidate solutions' function values. This renders those methods more robust as they are invariant to strictly increasing transformations of the objective function but at the same time the theoretical analysis becomes more difficult as we cannot exploit a well defined framework using (strong) properties of the function like convexity or smoothness.
Additionally, adaptive stochastic optimization algorithms typically have a complex state space which encodes the parameters of a probability distribution (e.g. mean and covariance matrix of a Gaussian vector) and other state vectors. This statespace is a manifold. While the algorithms are Markov chains, the complexity of the statespace makes that standard Markov chain theory tools do not directly apply. The same holds with tools stemming from stochastic approximation theory or Ordinary Differential Equation (ODE) theory where it is usually assumed that the underlying ODE (obtained by proper averaging and limit for learning rate to zero) has its critical points inside the search space. In contrast, in the cases we are interested, the critical points of the ODEs are at the boundary of the domain.
Last, since we aim at developing theory that one the one hand allows to analyze the main properties of stateoftheart methods and on the other hand is useful for algorithm design, we need to be careful to not use simplifications that would allow a proof to be done but would not capture the important properties of the algorithms. With that respect one tricky point is to develop theory that accounts for invariance properties.
To face those specific challenges, we need to develop novel theoretical frameworks exploiting invariance properties and accounting for peculiar statespaces. Those frameworks should allow to analyze one of the core properties of adaptive stochastic methods, namely linear convergence on the widest possible class of functions.
We are planning on approaching the question of linear convergence from three different complementary angles, using three different frameworks:

the Markov chain framework where the convergence derives from the analysis of the stability of a normalized Markov chain existing on scalinginvariant functions for translation and scaleinvariant algorithms [15]. This framework allows for a fine analysis where the exact convergence rate can be given as an implicit function of the invariant measure of the normalized Markov chain. Yet it requires the objective function to be scalinginvariant. The stability analysis can be particularly tricky as the Markov chain that needs to be studied writes as ${\Phi}_{t+1}=F({\Phi}_{t},{W}_{t+1})$ where $\{{W}_{t}:t>0\}$ are independent identically distributed and $F$ is typically discontinuous because the algorithms studied are comparisonbased. This implies that practical tools for analyzing a standard property like irreducibility, that rely on investigating the stability of underlying deterministic control models [26], cannot be used. Additionally, the construction of a drift to prove ergodicity is particularly delicate when the state space includes a (normalized) covariance matrix as it is the case for analyzing the CMAES algorithm.

The stochastic approximation or ODE framework. Those are standard techniques to prove the convergence of stochastic algorithms when an algorithm can be expressed as a stochastic approximation of the solution of a mean field ODE [16], [17], [24]. What is specific and induces difficulties for the algorithms we aim at analyzing is the nonstandard statespace since the ODE variables correspond to the statevariables of the algorithm (e.g. ${\mathbb{R}}^{n}\times {\mathbb{R}}_{>0}$ for stepsize adaptive algorithms, ${\mathbb{R}}^{n}\times {\mathbb{R}}_{>0}\times {S}_{++}^{n}$ where ${S}_{++}^{n}$ denotes the set of positive definite matrices if a covariance matrix is additionally adapted). Consequently, the ODE can have many critical points at the boundary of its definition domain (e.g. all points corresponding to ${\sigma}_{t}=0$ are critical points of the ODE) which is not typical. Also we aim at proving linear convergence, for that it is crucial that the learning rate does not decrease to zero which is nonstandard in ODE method.

The direct framework where we construct a global Lyapunov function for the original algorithm from which we deduce bounds on the hitting time to reach an $\u03f5$ball of the optimum. For this framework as for the ODE framework, we expect that the class of functions where we can prove linear convergence are composite of $g\circ f$ where $f$ is differentiable and $g:\mathrm{Im}\left(f\right)\to \mathbb{R}$ is strictly increasing and that we can show convergence to a local minimum.
We expect those frameworks to be complementary in the sense that the assumptions required are different. Typically, the ODE framework should allow for proofs under the assumptions that learning rates are small enough while it is not needed for the Markov chain framework. Hence this latter framework captures better the real dynamics of the algorithm, yet under the assumption of scalinginvariance of the objective functions. By studying the different frameworks in parallel, we expect to gain synergies and possibly understand what is the most promising approach for solving the holy grail question of the linear convergence of CMAES.