In this work we study high-probability bounds for stochastic subgradient methods under heavy tailed noise in Hilbert spaces. In this setting the noise is only assumed to have finite variance as opposed to a sub-Gaussian distribution for which it is known that standard subgradient methods enjoy high-probability bounds. We analyzed a clipped version of the projected stochastic subgradient method, where subgradient estimates are truncated whenever they have large norms. We show that this clipping strategy leads both to optimal anytime and finite horizon bounds for general averaging schemes of the iterates. We also show an application of our proposal to the case of kernel methods which gives an efficient and fully implementable algorithm for statistical supervised learning problems. Preliminary experiments are shown to support the validity of the method.
Dettaglio pubblicazione
2024, SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, Pages 953-977 (volume: 6)
High Probability Bounds for Stochastic Subgradient Schemes with Heavy Tailed Noise (01a Articolo in rivista)
Angela Parletta Daniela, Paudice Andrea, Pontil Massimiliano, Salzo Saverio
keywords