首页> 外文期刊>Journal of machine learning research >Adaptive Minimax Regression Estimation over Sparse $ell_q$-Hulls
【24h】

Adaptive Minimax Regression Estimation over Sparse $ell_q$-Hulls

机译:稀疏$ ell_q $ -Hulls上的自适应Minimax回归估计

获取原文
       

摘要

Given a dictionary of $M_n$ predictors, in a random designregression setting with $n$ observations, we constructestimators that target the best performance among all the linearcombinations of the predictors under a sparse $ell_q$-norm ($0leq q leq 1$) constraint on the linear coefficients. Besidesidentifying the optimal rates of convergence, our universalaggregation strategies by model mixing achieve the optimal ratessimultaneously over the full range of $0leq q leq 1$ for any$M_n$ and without knowledge of the $ell_q$-norm of the bestlinear coefficients to represent the regression function. Toallow model misspecification, our upper bound results areobtained in a framework of aggregation of estimates. A strikingfeature is that no specific relationship among the predictors isneeded to achieve the upper rates of convergence (hencepermitting basically arbitrary correlations between thepredictors). Therefore, whatever the true regression function(assumed to be uniformly bounded), our estimators automaticallyexploit any sparse representation of the regression function (ifany), to the best extent possible within the$ell_q$-constrained linear combinations for any $0 leq q leq1$. A sparse approximation result in the $ell_q$-hulls turnsout to be crucial to adaptively achieve minimax rate optimalaggregation. It precisely characterizes the number of termsneeded to achieve a prescribed accuracy of approximation to thebest linear combination in an $ell_q$-hull for $0 leq q leq1$. It offers the insight that the minimax rate of$ell_q$-aggregation is basically determined by an effectivemodel size, which is a sparsity index that depends on $q$,$M_n$, $n$, and the $ell_q$-norm bound in an easilyinterpretable way based on a classical model selection theorythat deals with a large number of models. color="gray">
机译:给定$ M_n $个预测变量的字典,在带有$ n $个观测值的随机设计回归设置中,我们构造一个估计量,针对稀疏的$ ell_q $-范数($ 0leq q leq 1 $)下所有预测变量线性组合中的最佳性能线性系数的约束。除了确定最优收敛速度,我们的通用聚集策略通过模型混合在任何$ M_n $的$ 0leq q leq 1 $整个范围内同时达到了最优速率,并且不知道最佳线性系数的$ ell_q $-范数来表示回归函数。为了避免模型错误指定,我们在估计汇总框架内获得了上限结果。一个显着的特征是,不需要预测变量之间的特定关系来实现较高的收敛速度(因此,允许预测变量之间基本上具有任意相关性)。因此,无论真实的回归函数是什么(假设是一致有界的),我们的估计器都会自动利用回归函数的任何稀疏表示(如果有的话),并尽可能在$ ell_q $约束的线性组合中,对于任何$ 0 leq q leq1 $ 。 $ ell_q $ -hulls投票结果中的稀疏近似结果对于自适应地实现minimax速率最优聚合至关重要。它精确地描述了为$ 0 leq q leq1 $获得$ ell_q $-船体中达到最佳线性组合近似规定精度所需的项数。它提供的见解是,$ ell_q $聚集的最小最大速率基本上由有效模型大小决定,有效模型大小是一个稀疏指数,取决于$ q $,$ M_n $,$ n $和$ ell_q $范数边界以经典的模型选择理论为基础,以易于解释的方式处理大量模型。 color =“ gray”>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号