Orateur
Description
In this talk, we show novel optimal (or near optimal) convergence rates for a clipped version of the projected stochastic subgradient method. We consider nonsmooth convex problems in Hilbert spaces over possibly unbounded domains, under heavy-tailed noise that possesses only the first $p$ moments for $p \in \left]1,2\right]$. For the last iterate, we establish convergence in expectation with rates of order $(\log^{1/p} k)/k^{(p-1)/p}$ and $1/k^{(p-1)/p}$ for infinite and finite-horizon respectively. We also derive new convergence rates, in expectation and with high probability, for the average iterate --- improving the state of the art. Those results are applied to the problem of supervised learning with kernels demonstrating the effectiveness of our theory. Finally, we give preliminary experiments.