28 juillet 2025 à 1 août 2025
Fuseau horaire Europe/Paris

Score Learning under the Manifold Hypothesis: Theory and Implications for Data Science

31 juil. 2025, 10:45
30m
F202

F202

Invited talk Machine learning ML

Orateur

Ya-Ping Hsieh (ETH Zürich)

Description

The score function plays a central role in modern generative modeling, particularly in diffusion models and related score-based methods. Despite its theoretical appeal, learning the score function in practice is notoriously difficult: it is sensitive to hyperparameter choices and prone to various forms of instability that often require ad hoc corrections.

In this talk, we study score learning under the popular manifold hypothesis, which posits that high-dimensional data concentrates near a low-dimensional manifold. A key and perhaps surprising insight is that, under this hypothesis, it is significantly easier to learn the underlying manifold structure than the full data distribution. We provide theoretical support for this claim and explore its consequences for score estimation and generative modeling. In particular, we show how this perspective can clarify challenges in training diffusion models and guide the development of more stable and efficient algorithms.

Author

Ya-Ping Hsieh (ETH Zürich)

Documents de présentation

Aucun document.