28 juillet 2025 à 1 août 2025
Fuseau horaire Europe/Paris

An stochastic differential equation perspective on stochastic convex optimization

31 juil. 2025, 10:45
30m
F102

F102

Contributed talk Stochastic Programming Stochastic Programming

Orateur

Rodrigo Maulen Soto

Description

We analyze the global and local behavior of gradient-like flows under stochastic errors towards the aim of solving convex optimization problems with noisy gradient input. We first study the unconstrained differentiable convex case, using a stochastic differential equation where the drift term is minus the gradient of the objective function and the diffusion term is either bounded or square-integrable. In this context, under Lipschitz continuity of the gradient, our first main result shows almost sure convergence of the objective and the trajectory process towards a minimizer of the objective function. We also provide a comprehensive complexity analysis by establishing several new pointwise and ergodic convergence rates in expectation for the convex, strongly convex, and (local) Łojasiewicz case. The latter, which involves local analysis, is challenging and requires non-trivial arguments from measure theory. Then, we extend our study to the constrained case and more generally to certain nonsmooth situations. We show that several of our results have natural extensions obtained by replacing the gradient of the objective function by a cocoercive monotone operator. This makes it possible to obtain similar convergence results for optimization problems with an additively "smooth + non-smooth" convex structure. Finally, we consider another extension of our results to non-smooth optimization which is based on the Moreau envelope.

Author

Co-auteurs

Prof. Hedy Attouch (Univ Montpellier) Prof. Jalal Fadili (ENSICAEN)

Documents de présentation

Aucun document.