Research

Paper

AI LLM February 20, 2026

Analyzing and Improving Chain-of-Thought Monitorability Through Information Theory

Authors

Usman Anwar, Tim Bakker, Dana Kianfar, Cristina Pinneri, Christos Louizos

Abstract

Chain-of-thought (CoT) monitors are LLM-based systems that analyze reasoning traces to detect when outputs may exhibit attributes of interest, such as test-hacking behavior during code generation. In this paper, we use information-theoretic analysis to show that non-zero mutual information between CoT and output is a necessary but not sufficient condition for CoT monitorability. We identify two sources of approximation error that may undermine the performance of CoT monitors in practice: information gap, which measures the extent to which the monitor can extract the information available in CoT, and elicitation error, which measures the extent to which the monitor approximates the optimal monitoring function. We further demonstrate that CoT monitorability can be systematically improved through targeted training objectives. To this end, we propose two complementary approaches: (a) an oracle-based method that directly rewards the monitored model for producing CoTs that maximize monitor accuracy, and (b) a more practical, label-free approach that maximizes conditional mutual information between outputs and CoTs. Across multiple different environments, we show both methods significantly improve monitor accuracy while preventing CoT degeneration even when training against a monitor, thereby mitigating reward hacking when the task reward is imperfectly specified.

Metadata

arXiv ID: 2602.18297
Provider: ARXIV
Primary Category: cs.LG
Published: 2026-02-20
Fetched: 2026-02-23 05:33

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.18297v1</id>\n    <title>Analyzing and Improving Chain-of-Thought Monitorability Through Information Theory</title>\n    <updated>2026-02-20T15:50:30Z</updated>\n    <link href='https://arxiv.org/abs/2602.18297v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.18297v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Chain-of-thought (CoT) monitors are LLM-based systems that analyze reasoning traces to detect when outputs may exhibit attributes of interest, such as test-hacking behavior during code generation. In this paper, we use information-theoretic analysis to show that non-zero mutual information between CoT and output is a necessary but not sufficient condition for CoT monitorability. We identify two sources of approximation error that may undermine the performance of CoT monitors in practice: information gap, which measures the extent to which the monitor can extract the information available in CoT, and elicitation error, which measures the extent to which the monitor approximates the optimal monitoring function. We further demonstrate that CoT monitorability can be systematically improved through targeted training objectives. To this end, we propose two complementary approaches: (a) an oracle-based method that directly rewards the monitored model for producing CoTs that maximize monitor accuracy, and (b) a more practical, label-free approach that maximizes conditional mutual information between outputs and CoTs. Across multiple different environments, we show both methods significantly improve monitor accuracy while preventing CoT degeneration even when training against a monitor, thereby mitigating reward hacking when the task reward is imperfectly specified.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.AI'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CL'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.IT'/>\n    <published>2026-02-20T15:50:30Z</published>\n    <arxiv:comment>First two authors contributed equally</arxiv:comment>\n    <arxiv:primary_category term='cs.LG'/>\n    <author>\n      <name>Usman Anwar</name>\n    </author>\n    <author>\n      <name>Tim Bakker</name>\n    </author>\n    <author>\n      <name>Dana Kianfar</name>\n    </author>\n    <author>\n      <name>Cristina Pinneri</name>\n    </author>\n    <author>\n      <name>Christos Louizos</name>\n    </author>\n  </entry>"
}