Research

Paper

AI LLM March 09, 2026

Leaderboard Incentives: Model Rankings under Strategic Post-Training

Authors

Yatong Chen, Guanhua Zhang, Moritz Hardt

Abstract

Influential benchmarks incentivize competing model developers to strategically allocate post-training resources toward improvements on the leaderboard, a phenomenon dubbed benchmaxxing or training on the test task. In this work, we initiate a principled study of the incentive structure that benchmarks induce. We model benchmarking as a Stackelberg game between a benchmark designer who chooses an evaluation protocol and multiple model developers who compete simultaneously in a subgame given by the designer's choice. Each competitor has a model of unknown latent quality and can inflate its observed score by allocating resources to benchmark-specific improvements. First, we prove that current benchmarks induce games for which no Nash equilibrium between model developers exists. This result suggests one explanation for why current practice leads to misaligned incentives, prompting model developers to strategize in opaque ways. However, we prove that under mild conditions, a recently proposed evaluation protocol, called tune-before-test, induces a benchmark with a unique Nash equilibrium that ranks models by latent quality. This positive result demonstrates that benchmarks need not set bad incentives, even if current evaluations do.

Metadata

arXiv ID: 2603.08371
Provider: ARXIV
Primary Category: cs.GT
Published: 2026-03-09
Fetched: 2026-03-10 05:43

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.08371v1</id>\n    <title>Leaderboard Incentives: Model Rankings under Strategic Post-Training</title>\n    <updated>2026-03-09T13:33:20Z</updated>\n    <link href='https://arxiv.org/abs/2603.08371v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.08371v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Influential benchmarks incentivize competing model developers to strategically allocate post-training resources toward improvements on the leaderboard, a phenomenon dubbed benchmaxxing or training on the test task. In this work, we initiate a principled study of the incentive structure that benchmarks induce. We model benchmarking as a Stackelberg game between a benchmark designer who chooses an evaluation protocol and multiple model developers who compete simultaneously in a subgame given by the designer's choice. Each competitor has a model of unknown latent quality and can inflate its observed score by allocating resources to benchmark-specific improvements. First, we prove that current benchmarks induce games for which no Nash equilibrium between model developers exists. This result suggests one explanation for why current practice leads to misaligned incentives, prompting model developers to strategize in opaque ways. However, we prove that under mild conditions, a recently proposed evaluation protocol, called tune-before-test, induces a benchmark with a unique Nash equilibrium that ranks models by latent quality. This positive result demonstrates that benchmarks need not set bad incentives, even if current evaluations do.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.GT'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <published>2026-03-09T13:33:20Z</published>\n    <arxiv:primary_category term='cs.GT'/>\n    <author>\n      <name>Yatong Chen</name>\n    </author>\n    <author>\n      <name>Guanhua Zhang</name>\n    </author>\n    <author>\n      <name>Moritz Hardt</name>\n    </author>\n  </entry>"
}