2–8 Nov 2025
TIFR Mumbai
Asia/Kolkata timezone

Robustness and over-parameterization

3 Nov 2025, 18:00
20m
AG69

AG69

Speaker

Piyush Srivastava (Tata Institute of Fundamental Research)

Description

In many modern machine learning applications, models are often trained
to near zero "training loss" (in other words, to interpolate the
training data), while also having far more training parameters than the
"number of data points". This appears to violate traditional
rules-of-thumb for avoiding overfitting, and considerable work has thus
been devoted to gain a better understanding of such
over-parameterization. A more recent development is an interesting
direction suggested by Bubeck and Sellke, who postulated that, in
certain settings, a much larger number of parameters may indeed be
required to interpolate the training data if the model being trained is
also required to be "robust": small modifications to the model input
should not lead to very large changes in the model output.

In this talk, we will survey this formulation of the connections between
robustness and over-parameterization. We will also present the
conceptual view that a bias-variance type decomposition for the loss
function lies at the heart of the results of Bubeck and Sellke, and then
use this idea to show that the setting of losses corresponding to
Bregman divergences is the natural setting for understanding the
connection between robustness and over-parameterization in this formulation.

Parallel Session (for talks only) Algorithms and artificial intelligence

Authors

Jatin Batra (TIFR Mumbai) Piyush Srivastava (Tata Institute of Fundamental Research) Santanu Das (TIFR Mumbai)

Presentation materials

There are no materials yet.