Research

Paper

AI LLM March 23, 2026

INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation

Authors

Alexandra Bazarova, Andrei Volodichev, Daria Kotova, Alexey Zaytsev

Abstract

While retrieval-augmented generation (RAG) significantly improves the factual reliability of LLMs, it does not eliminate hallucinations, so robust uncertainty quantification (UQ) remains essential. In this paper, we reveal that standard entropy-based UQ methods often fail in RAG settings due to a mechanistic paradox. An internal "tug-of-war" inherent to context utilization appears: while induction heads promote grounded responses by copying the correct answer, they collaterally trigger the previously established "entropy neurons". This interaction inflates predictive entropy, causing the model to signal false uncertainty on accurate outputs. To address this, we propose INTRYGUE (Induction-Aware Entropy Gating for Uncertainty Estimation), a mechanistically grounded method that gates predictive entropy based on the activation patterns of induction heads. Evaluated across four RAG benchmarks and six open-source LLMs (4B to 13B parameters), INTRYGUE consistently matches or outperforms a wide range of UQ baselines. Our findings demonstrate that hallucination detection in RAG benefits from combining predictive uncertainty with interpretable, internal signals of context utilization.

Metadata

arXiv ID: 2603.21607
Provider: ARXIV
Primary Category: cs.AI
Published: 2026-03-23
Fetched: 2026-03-24 06:02

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.21607v1</id>\n    <title>INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation</title>\n    <updated>2026-03-23T06:02:16Z</updated>\n    <link href='https://arxiv.org/abs/2603.21607v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.21607v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>While retrieval-augmented generation (RAG) significantly improves the factual reliability of LLMs, it does not eliminate hallucinations, so robust uncertainty quantification (UQ) remains essential. In this paper, we reveal that standard entropy-based UQ methods often fail in RAG settings due to a mechanistic paradox. An internal \"tug-of-war\" inherent to context utilization appears: while induction heads promote grounded responses by copying the correct answer, they collaterally trigger the previously established \"entropy neurons\". This interaction inflates predictive entropy, causing the model to signal false uncertainty on accurate outputs. To address this, we propose INTRYGUE (Induction-Aware Entropy Gating for Uncertainty Estimation), a mechanistically grounded method that gates predictive entropy based on the activation patterns of induction heads. Evaluated across four RAG benchmarks and six open-source LLMs (4B to 13B parameters), INTRYGUE consistently matches or outperforms a wide range of UQ baselines. Our findings demonstrate that hallucination detection in RAG benefits from combining predictive uncertainty with interpretable, internal signals of context utilization.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.AI'/>\n    <published>2026-03-23T06:02:16Z</published>\n    <arxiv:primary_category term='cs.AI'/>\n    <author>\n      <name>Alexandra Bazarova</name>\n    </author>\n    <author>\n      <name>Andrei Volodichev</name>\n    </author>\n    <author>\n      <name>Daria Kotova</name>\n    </author>\n    <author>\n      <name>Alexey Zaytsev</name>\n    </author>\n  </entry>"
}