Research

Paper

TESTING February 26, 2026

Asymptotic Theory and Sequential Test for General Multi-Armed Bandit Process

Authors

Li Yang, Xiaodong Yan, Dandan Jiang

Abstract

Multi-armed bandit (MAB) processes constitute a foundational subclass of reinforcement learning problems and represent a central topic in statistical decision theory, but are limited to simultaneous adaptive allocation and sequential test, because of the absence of asymptotic theory under non-i.i.d sequence and sublinear information. To address this open challenge, we propose Urn Bandit (UNB) process to integrate the reinforcement mechanism of urn probabilistic models with MAB principles, ensuring almost sure convergence of resource allocation to optimal arms. We establish the joint functional central limit theorem (FCLT) for consistent estimators of expected rewards under non-i.i.d., non-sub-Gaussian and sublinear reward samples with pairwise correlations across arms. To overcome the limitations of existing methods that focus mainly on cumulative regret, we establish the asymptotic theory along with adaptive allocation that serves powerful sequential test, such as arms comparison, A/B testing, and policy valuation. Simulation studies and real data analysis demonstrate that UNB maintains statistical test performance of equal randomization (ER) design but obtain more average rewards like classical MAB processes.

Metadata

arXiv ID: 2602.22768
Provider: ARXIV
Primary Category: stat.ME
Published: 2026-02-26
Fetched: 2026-02-27 04:35

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.22768v1</id>\n    <title>Asymptotic Theory and Sequential Test for General Multi-Armed Bandit Process</title>\n    <updated>2026-02-26T08:59:20Z</updated>\n    <link href='https://arxiv.org/abs/2602.22768v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.22768v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Multi-armed bandit (MAB) processes constitute a foundational subclass of reinforcement learning problems and represent a central topic in statistical decision theory, but are limited to simultaneous adaptive allocation and sequential test, because of the absence of asymptotic theory under non-i.i.d sequence and sublinear information. To address this open challenge, we propose Urn Bandit (UNB) process to integrate the reinforcement mechanism of urn probabilistic models with MAB principles, ensuring almost sure convergence of resource allocation to optimal arms. We establish the joint functional central limit theorem (FCLT) for consistent estimators of expected rewards under non-i.i.d., non-sub-Gaussian and sublinear reward samples with pairwise correlations across arms. To overcome the limitations of existing methods that focus mainly on cumulative regret, we establish the asymptotic theory along with adaptive allocation that serves powerful sequential test, such as arms comparison, A/B testing, and policy valuation. Simulation studies and real data analysis demonstrate that UNB maintains statistical test performance of equal randomization (ER) design but obtain more average rewards like classical MAB processes.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='stat.ME'/>\n    <published>2026-02-26T08:59:20Z</published>\n    <arxiv:primary_category term='stat.ME'/>\n    <author>\n      <name>Li Yang</name>\n    </author>\n    <author>\n      <name>Xiaodong Yan</name>\n    </author>\n    <author>\n      <name>Dandan Jiang</name>\n    </author>\n  </entry>"
}