Research

Paper

AI LLM February 24, 2026

Mitigating Preference Leakage via Strict Estimator Separation for Normative Generative Ranking

Authors

Dalia Nahhas, Xiaohao Cai, Imran Razzak, Shoaib Jameel

Abstract

In Generative Information Retrieval (GenIR), the bottleneck has shifted from generation to the selection of candidates, particularly for normative criteria such as cultural relevance. Current LLM-as-a-Judge evaluations often suffer from circularity and preference leakage, where overlapping supervision and evaluation models inflate performance. We address this by formalising cultural relevance as a within-query ranking task and introducing a leakage-free two-judge framework that strictly separates supervision (Judge B) from evaluation (Judge A). On a new benchmark of 33,052 (NGR-33k) culturally grounded stories, we find that while classical baselines yield only modest gains, a dense bi-encoder distilled from a Judge-B-supervised Cross-Encoder is highly effective. Although the Cross-Encoder provides a strong supervision signal for distillation, the distilled BGE-M3 model substantially outperforms it under leakage-free Judge~A evaluation. We validate our framework on the human-curated Moral Stories dataset, showing strong alignment with human norms. Our results demonstrate that rigorous evaluator separation is a prerequisite for credible GenIR evaluation, proving that subtle cultural preferences can be distilled into efficient rankers without leakage.

Metadata

arXiv ID: 2602.20800
Provider: ARXIV
Primary Category: cs.IR
Published: 2026-02-24
Fetched: 2026-02-25 06:05

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.20800v1</id>\n    <title>Mitigating Preference Leakage via Strict Estimator Separation for Normative Generative Ranking</title>\n    <updated>2026-02-24T11:38:36Z</updated>\n    <link href='https://arxiv.org/abs/2602.20800v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.20800v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>In Generative Information Retrieval (GenIR), the bottleneck has shifted from generation to the selection of candidates, particularly for normative criteria such as cultural relevance. Current LLM-as-a-Judge evaluations often suffer from circularity and preference leakage, where overlapping supervision and evaluation models inflate performance. We address this by formalising cultural relevance as a within-query ranking task and introducing a leakage-free two-judge framework that strictly separates supervision (Judge B) from evaluation (Judge A). On a new benchmark of 33,052 (NGR-33k) culturally grounded stories, we find that while classical baselines yield only modest gains, a dense bi-encoder distilled from a Judge-B-supervised Cross-Encoder is highly effective. Although the Cross-Encoder provides a strong supervision signal for distillation, the distilled BGE-M3 model substantially outperforms it under leakage-free Judge~A evaluation. We validate our framework on the human-curated Moral Stories dataset, showing strong alignment with human norms. Our results demonstrate that rigorous evaluator separation is a prerequisite for credible GenIR evaluation, proving that subtle cultural preferences can be distilled into efficient rankers without leakage.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.IR'/>\n    <published>2026-02-24T11:38:36Z</published>\n    <arxiv:primary_category term='cs.IR'/>\n    <author>\n      <name>Dalia Nahhas</name>\n    </author>\n    <author>\n      <name>Xiaohao Cai</name>\n    </author>\n    <author>\n      <name>Imran Razzak</name>\n    </author>\n    <author>\n      <name>Shoaib Jameel</name>\n    </author>\n  </entry>"
}