Research

Paper

TESTING February 23, 2026

Model Selection in High-Dimensional Linear Regression using Boosting with Multiple Testing

Authors

George Kapetanios, Vasilis Sarafidis, Alexia Ventouri

Abstract

High-dimensional regression specification and analysis is a complex and active area of research in statistics, machine learning, and econometrics. This paper proposes a new approach, Boosting with Multiple Testing (BMT), which combines forward stepwise variable selection with the multiple testing framework of Chudik et al (2018). At each stage, the model is updated by adding only the most significant regressor conditional on those already included, while a family-wise multiple testing filter is applied to the remaining candidates. In this way, the method retains the strong screening properties of Chudik et al (2018) while operating in a less greedy manner with respect to proxy and noise variables. Using sharp probability inequalities for heterogeneous strongly mixing processes from Dendramis et al (2022), we show that BMT enjoys oracle type properties relative to an approximating model that includes all true signals and excludes pure noise variables: this model is selected with probability tending to one, and the resulting estimator achieves standard parametric rates for prediction error and coefficient estimation. Additional results establish conditions under which BMT recovers the exact true model and avoids selection of proxy signals. Monte Carlo experiments indicate that BMT performs very well relative to OCMT and Lasso type procedures, delivering higher model selection accuracy and smaller RMSE for the estimated coefficients, especially under strong multicollinearity of the regressors. Two empirical illustrations based on a large set of macro-financial indicators as covariates, show that BMT yields sparse, interpretable specifications with favourable out-of-sample performance.

Metadata

arXiv ID: 2602.19705
Provider: ARXIV
Primary Category: econ.EM
Published: 2026-02-23
Fetched: 2026-02-24 04:38

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.19705v1</id>\n    <title>Model Selection in High-Dimensional Linear Regression using Boosting with Multiple Testing</title>\n    <updated>2026-02-23T10:55:18Z</updated>\n    <link href='https://arxiv.org/abs/2602.19705v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.19705v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>High-dimensional regression specification and analysis is a complex and active area of research in statistics, machine learning, and econometrics. This paper proposes a new approach, Boosting with Multiple Testing (BMT), which combines forward stepwise variable selection with the multiple testing framework of Chudik et al (2018). At each stage, the model is updated by adding only the most significant regressor conditional on those already included, while a family-wise multiple testing filter is applied to the remaining candidates. In this way, the method retains the strong screening properties of Chudik et al (2018) while operating in a less greedy manner with respect to proxy and noise variables. Using sharp probability inequalities for heterogeneous strongly mixing processes from Dendramis et al (2022), we show that BMT enjoys oracle type properties relative to an approximating model that includes all true signals and excludes pure noise variables: this model is selected with probability tending to one, and the resulting estimator achieves standard parametric rates for prediction error and coefficient estimation. Additional results establish conditions under which BMT recovers the exact true model and avoids selection of proxy signals. Monte Carlo experiments indicate that BMT performs very well relative to OCMT and Lasso type procedures, delivering higher model selection accuracy and smaller RMSE for the estimated coefficients, especially under strong multicollinearity of the regressors. Two empirical illustrations based on a large set of macro-financial indicators as covariates, show that BMT yields sparse, interpretable specifications with favourable out-of-sample performance.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='econ.EM'/>\n    <published>2026-02-23T10:55:18Z</published>\n    <arxiv:primary_category term='econ.EM'/>\n    <author>\n      <name>George Kapetanios</name>\n    </author>\n    <author>\n      <name>Vasilis Sarafidis</name>\n    </author>\n    <author>\n      <name>Alexia Ventouri</name>\n    </author>\n  </entry>"
}