首页> 外文期刊>Electronic Journal of Statistics >Randomized maximum-contrast selection: Subagging for large-scale regression
【24h】

Randomized maximum-contrast selection: Subagging for large-scale regression

机译:随机化最大对比度选择:细分用于大规模回归

获取原文
       

摘要

We introduce a general method for variable selection in a large-scale regression setting where both the number of parameters and the number of samples are extremely large. The proposed method is based on careful combination of penalized estimators, each applied to a random projection of the sample space into a low-dimensional space. In one special case that we study in detail, the random projections are divided into non-overlapping blocks, each consisting of only a small portion of the original data. Within each block we select the projection yielding the smallest out-of-sample error. Our random ensemble estimator then aggregates the results according to a new maximal-contrast voting scheme to determine the final selected set. Our theoretical results illustrate the effect on performance of increasing the number of non-overlapping blocks. Moreover, we demonstrate that statistical optimality is retained along with the computational speedup. The proposed method achieves minimax rates for approximate recovery over all estimators, using the full set of samples. Furthermore, our theoretical results allow the number of subsamples to grow with the subsample size and do not require irrepresentable condition. The estimator is also compared empirically with several other popular high-dimensional estimators via an extensive simulation study, which reveals its excellent finite-sample performance.
机译:我们介绍了在参数数量和样本数量都非常大的大规模回归设置中选择变量的通用方法。所提出的方法是基于精心组合的惩罚估计量,每个估计量都应用于样本空间到低维空间的随机投影。在我们将详细研究的一种特殊情况下,随机投影被划分为不重叠的块,每个块仅包含原始数据的一小部分。在每个块中,我们选择产生最小样本外误差的投影。然后,我们的随机整体估算器会根据新的最大对比度投票方案汇总结果,以确定最终选择的集合。我们的理论结果说明了增加不重叠块数量对性能的影响。此外,我们证明了统计最优性随着计算速度的提高而得以保留。所提出的方法使用全部样本,可在所有估计量上实现近似恢复的minimax速率。此外,我们的理论结果允许子样本的数量随子样本的大小而增长,并且不需要无可辩驳的条件。通过广泛的仿真研究,该估算器还与其他几种流行的高维估算器进行了经验比较,从而揭示了其出色的有限样本性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号