7
0

Revisiting Randomization in Greedy Model Search

Xin Chen
Jason M. Klusowski
Yan Shuo Tan
Chang Yu
Author Contacts:
Main:17 Pages
21 Figures
Bibliography:5 Pages
Appendix:32 Pages
Abstract

Combining randomized estimators in an ensemble, such as via random forests, has become a fundamental technique in modern data science, but can be computationally expensive. Furthermore, the mechanism by which this improves predictive performance is poorly understood. We address these issues in the context of sparse linear regression by proposing and analyzing an ensemble of greedy forward selection estimators that are randomized by feature subsampling -- at each iteration, the best feature is selected from within a random subset. We design a novel implementation based on dynamic programming that greatly improves its computational efficiency. Furthermore, we show via careful numerical experiments that our method can outperform popular methods such as lasso and elastic net across a wide range of settings. Next, contrary to prevailing belief that randomized ensembling is analogous to shrinkage, we show via numerical experiments that it can simultaneously reduce training error and degrees of freedom, thereby shifting the entire bias-variance trade-off curve of the base estimator. We prove this fact rigorously in the setting of orthogonal features, in which case, the ensemble estimator rescales the ordinary least squares coefficients with a two-parameter family of logistic weights, thereby enlarging the model search space. These results enhance our understanding of random forests and suggest that implicit regularization in general may have more complicated effects than explicit regularization.

View on arXiv
@article{chen2025_2506.15643,
  title={ Revisiting Randomization in Greedy Model Search },
  author={ Xin Chen and Jason M. Klusowski and Yan Shuo Tan and Chang Yu },
  journal={arXiv preprint arXiv:2506.15643},
  year={ 2025 }
}
Comments on this paper