28 juillet 2025 à 1 août 2025
Fuseau horaire Europe/Paris

Global convergence of stochastic gradient bandits for any learning rates

Non programmé
30m
F206

F206

Invited talk Machine learning ML

Orateur

Jincheng Mei (Google DeepMind)

Description

We provide a new understanding of the stochastic gradient bandit algorithm by showing that it converges to a globally optimal policy almost surely using any constant learning rate. This result demonstrates that the stochastic gradient algorithm continues to balance exploration and exploitation appropriately even in scenarios where standard smoothness and noise control assumptions break down. The proofs are based on novel findings about action sampling rates and the relationship between cumulative progress and noise, and extend the current understanding of how simple stochastic gradient methods behave in bandit settings.

Author

Jincheng Mei (Google DeepMind)

Documents de présentation

Aucun document.