Research

Paper

TESTING March 02, 2026

Learning Structured Reasoning via Tractable Trajectory Control

Authors

Po-Nien Kung, Zhen Yang, Jeffrey Luo, Cheng-Fu Yang, Haikang Deng, Zi-Yi Dou, Yinfei Yang, Nanyun Peng, Zhe Gan, Kai-Wei Chang

Abstract

Large language models can exhibit emergent reasoning behaviors, often manifested as recurring lexical patterns (e.g., "wait," indicating verification). However, complex reasoning trajectories remain sparse in unconstrained sampling, and standard RL often fails to guarantee the acquisition of diverse reasoning behaviors. We propose a systematic discovery and reinforcement of diverse reasoning patterns through structured reasoning, a paradigm that requires targeted exploration of specific reasoning patterns during the RL process. To this end, we propose Ctrl-R, a framework for learning structured reasoning via tractable trajectory control that actively guides the rollout process, incentivizing the exploration of diverse reasoning patterns that are critical for complex problem-solving. The resulting behavior policy enables accurate importance-sampling estimation, supporting unbiased on-policy optimization. We further introduce a power-scaling factor on the importance-sampling weights, allowing the policy to selectively learn from exploratory, out-of-distribution trajectories while maintaining stable optimization. Experiments demonstrate that Ctrl-R enables effective exploration and internalization of previously unattainable reasoning patterns, yielding consistent improvements across language and vision-language models on mathematical reasoning tasks.

Metadata

arXiv ID: 2603.01641
Provider: ARXIV
Primary Category: cs.AI
Published: 2026-03-02
Fetched: 2026-03-03 04:34

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.01641v1</id>\n    <title>Learning Structured Reasoning via Tractable Trajectory Control</title>\n    <updated>2026-03-02T09:18:19Z</updated>\n    <link href='https://arxiv.org/abs/2603.01641v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.01641v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Large language models can exhibit emergent reasoning behaviors, often manifested as recurring lexical patterns (e.g., \"wait,\" indicating verification). However, complex reasoning trajectories remain sparse in unconstrained sampling, and standard RL often fails to guarantee the acquisition of diverse reasoning behaviors. We propose a systematic discovery and reinforcement of diverse reasoning patterns through structured reasoning, a paradigm that requires targeted exploration of specific reasoning patterns during the RL process. To this end, we propose Ctrl-R, a framework for learning structured reasoning via tractable trajectory control that actively guides the rollout process, incentivizing the exploration of diverse reasoning patterns that are critical for complex problem-solving. The resulting behavior policy enables accurate importance-sampling estimation, supporting unbiased on-policy optimization. We further introduce a power-scaling factor on the importance-sampling weights, allowing the policy to selectively learn from exploratory, out-of-distribution trajectories while maintaining stable optimization. Experiments demonstrate that Ctrl-R enables effective exploration and internalization of previously unattainable reasoning patterns, yielding consistent improvements across language and vision-language models on mathematical reasoning tasks.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.AI'/>\n    <published>2026-03-02T09:18:19Z</published>\n    <arxiv:primary_category term='cs.AI'/>\n    <author>\n      <name>Po-Nien Kung</name>\n    </author>\n    <author>\n      <name>Zhen Yang</name>\n    </author>\n    <author>\n      <name>Jeffrey Luo</name>\n    </author>\n    <author>\n      <name>Cheng-Fu Yang</name>\n    </author>\n    <author>\n      <name>Haikang Deng</name>\n    </author>\n    <author>\n      <name>Zi-Yi Dou</name>\n    </author>\n    <author>\n      <name>Yinfei Yang</name>\n    </author>\n    <author>\n      <name>Nanyun Peng</name>\n    </author>\n    <author>\n      <name>Zhe Gan</name>\n    </author>\n    <author>\n      <name>Kai-Wei Chang</name>\n    </author>\n  </entry>"
}