Version française / Séminaires
Séminaire MODAL'X : Rafal Kulil (Université d'Ottawa)
Publié le 19 septembre 2025
–
Mis à jour le 2 octobre 2025
Heavy tails in Machine Learning algorithms
Date(s)
le 20 novembre 2025
14h00 - 15h00
Lieu(x)
Résumé : Stochastic Gradient Descent (SGD) has become the most popular optimization algorithm used in Machine Learning. It was empirically observed that some characteristic of SGD exhibit heavy tails.
In this talk we will (partially) explain the mechanisms that lead to emergence of heavy tails in both online and offline SGDs.Furthermore, we will study SGD through the lenses of stochastic processes obtaining stable convergence. The main tools to prove relevant results are borrowed from the theory of heavy-tailed time series.
In this talk we will (partially) explain the mechanisms that lead to emergence of heavy tails in both online and offline SGDs.Furthermore, we will study SGD through the lenses of stochastic processes obtaining stable convergence. The main tools to prove relevant results are borrowed from the theory of heavy-tailed time series.
Mis à jour le 02 octobre 2025