Séminaire MODAL'X : Rafal Kulil (Université d'Ottawa)

Publié le 19 septembre 2025 Mis à jour le 2 octobre 2025

Heavy tails in Machine Learning algorithms

Date(s)

le 20 novembre 2025

14h00 - 15h00
Lieu(x)

Bâtiment Maurice Allais (G)

Entresol, salle Modal'X (E-27)
Plan d'accès
Résumé : Stochastic Gradient Descent (SGD) has become the most popular optimization algorithm used in Machine Learning. It was empirically observed that some characteristic of SGD exhibit heavy tails.
In this talk we will (partially) explain the mechanisms that lead to emergence of heavy tails in both online and offline SGDs.Furthermore, we will study SGD through the lenses of stochastic processes obtaining stable convergence. The main tools to prove relevant results are borrowed from the theory of heavy-tailed time series.

Mis à jour le 02 octobre 2025