An Introduction to Mean-Field Langevin Dynamics

An Introduction to Mean-Field Langevin Dynamics

Optimization over the space of probability measures is not only widely applicable, but also offers a useful perspective for analyzing certain complicated finite-dimensional nonconvex optimization problems. In particular, lifting such problems to optimization problems over probability measures can lead to better structural properties, such as convexity. Mean-field Langevin dynamics provides a representative example of this idea. Its central motivation is that some highly nonconvex optimization problems arising in neural network training become better behaved when reformulated as the optimization of a functional on the space of probability measures. This viewpoint also makes it possible to build a theoretical foundation for understanding the convergence of SGD. In what follows, we briefly introduce this perspective, mainly based on the paper by [Hu, Kaitong, et al]. The main analytical framework of this theory can be illustrated by Figure 2.

Read more
An Example of Gamma Convergence

An Example of Gamma Convergence

Let $\mathcal{P}_2(\mathbb{R}^d)$ denote the space of probability measures on $\mathbb{R}^d$ with finite second moment and $F: \mathcal{P}_2(\mathbb{R}^d)\to \mathbb{R}$ be the potential functional. We are interested in the following optimization problem

$$
\min_{m\in \mathcal{P}_2(\mathbb{R}^d)} F(m).
$$

Read more