Research

Paper

TESTING March 03, 2026

The Price of Robustness: Stable Classifiers Need Overparameterization

Authors

Jonas von Berg, Adalbert Fono, Massimiliano Datres, Sohir Maskey, Gitta Kutyniok

Abstract

The relationship between overparameterization, stability, and generalization remains incompletely understood in the setting of discontinuous classifiers. We address this gap by establishing a generalization bound for finite function classes that improves inversely with class stability, defined as the expected distance to the decision boundary in the input domain (margin). Interpreting class stability as a quantifiable notion of robustness, we derive as a corollary a law of robustness for classification that extends the results of Bubeck and Sellke beyond smoothness assumptions to discontinuous functions. In particular, any interpolating model with $p \approx n$ parameters on $n$ data points must be unstable, implying that substantial overparameterization is necessary to achieve high stability. We obtain analogous results for parameterized infinite function classes by analyzing a stronger robustness measure derived from the margin in the codomain, which we refer to as the normalized co-stability. Experiments support our theory: stability increases with model size and correlates with test performance, while traditional norm-based measures remain largely uninformative.

Metadata

arXiv ID: 2603.02806
Provider: ARXIV
Primary Category: cs.LG
Published: 2026-03-03
Fetched: 2026-03-04 03:41

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.02806v1</id>\n    <title>The Price of Robustness: Stable Classifiers Need Overparameterization</title>\n    <updated>2026-03-03T09:47:06Z</updated>\n    <link href='https://arxiv.org/abs/2603.02806v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.02806v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>The relationship between overparameterization, stability, and generalization remains incompletely understood in the setting of discontinuous classifiers. We address this gap by establishing a generalization bound for finite function classes that improves inversely with class stability, defined as the expected distance to the decision boundary in the input domain (margin). Interpreting class stability as a quantifiable notion of robustness, we derive as a corollary a law of robustness for classification that extends the results of Bubeck and Sellke beyond smoothness assumptions to discontinuous functions. In particular, any interpolating model with $p \\approx n$ parameters on $n$ data points must be unstable, implying that substantial overparameterization is necessary to achieve high stability. We obtain analogous results for parameterized infinite function classes by analyzing a stronger robustness measure derived from the margin in the codomain, which we refer to as the normalized co-stability. Experiments support our theory: stability increases with model size and correlates with test performance, while traditional norm-based measures remain largely uninformative.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <published>2026-03-03T09:47:06Z</published>\n    <arxiv:comment>29 pages, 9 figures. Accepted at ICLR 2026</arxiv:comment>\n    <arxiv:primary_category term='cs.LG'/>\n    <arxiv:journal_ref>In Proceedings of the Fourteenth International Conference on Learning Representations (ICLR), 2026</arxiv:journal_ref>\n    <author>\n      <name>Jonas von Berg</name>\n    </author>\n    <author>\n      <name>Adalbert Fono</name>\n    </author>\n    <author>\n      <name>Massimiliano Datres</name>\n    </author>\n    <author>\n      <name>Sohir Maskey</name>\n    </author>\n    <author>\n      <name>Gitta Kutyniok</name>\n    </author>\n  </entry>"
}