首页> 外文学位 >Simultaneous variable selection and simultaneous subspace selection for multitask learning.
【24h】

Simultaneous variable selection and simultaneous subspace selection for multitask learning.

机译:用于多任务学习的同时变量选择和同时子空间选择。

获取原文
获取原文并翻译 | 示例

摘要

This thesis considers the problem of simultaneous covariate selection and simultaneous subspace selection for a group of learning problems, e.g. regression or classification problems, which are defined over the same covariate space and which are assumed "related" in the sense that a small number of covariate or respectively a small dimensional subspace contains the information relevant to all learning problems. We use a ℓ1/ℓ2 block-regularization scheme that groups coefficients associated with each covariate across different classification problems, so that similar sparsity patterns in all models are encouraged. We propose a blockwise path-following scheme that approximately traces the regularization path and which takes advantage computationally of the sparsity of solutions at high regularization levels.;We then show how to use random projections to extend this approach to the problem of joint subspace selection, where multiple predictors are found in a common low-dimensional subspace and show that our algorithmically efficient scheme approximates the regularization by the trace norm.;Finally, in the context of K-dimensional multivariate linear regression, we study, from a theoretical point of view, the recovery of the union support---defined as the set of covariates relevant to at least one of the K output predictions---using the proposed ℓ1/ℓ2 regularization scheme. We show that the statistical complexity of the problem, measured by the number of observations needed to recover the correct union support with high probability, depends on a function that we introduce and analyze: the sparsity-overlap function. The theory, in accordance with empirical simulations, characterizes the situations in which selecting variables simultaneously based on our scheme is more efficient than selecting them separately.
机译:本论文考虑了一组学习问题的同时协变量选择和同时子空间选择的问题。回归或分类问题,它们定义在相同的协变量空间上,并在少量协变量或小维子空间包含与所有学习问题相关的信息的意义上被假定为“相关”。我们使用一个ℓ 1 /ℓ 2块正则化方案,该方案将与每个协变量相关联的系数归为不同的分类问题,从而鼓励在所有模型中使用相似的稀疏模式。我们提出了一种分块路径跟踪方案,该方案近似跟踪正则化路径,并在高正则化水平上利用了解决方案稀疏性的计算优势;然后我们展示了如何使用随机投影将该方法扩展到联合子空间选择问题,在一个常见的低维子空间中找到多个预测变量,并表明我们的算法有效方案通过跟踪范数逼近了正则化;最后,在K维多元线性回归的背景下,我们从理论的角度进行了研究,联合提议的恢复-被定义为与K个输出预测中至少一个相关的协变量集合-使用提议的ℓ 1 /ℓ 2正则化方案进行恢复。我们表明,问题的统计复杂性由以高概率恢复正确的工会支持所需的观察次数来衡量,取决于我们引入和分析的函数:稀疏重叠函数。根据经验模拟,该理论描述了一种情况,在这种情况下,根据我们的方案同时选择变量比单独选择变量更有效。

著录项

  • 作者单位

    University of California, Berkeley.;

  • 授予单位 University of California, Berkeley.;
  • 学科 Statistics.;Artificial Intelligence.
  • 学位 Ph.D.
  • 年度 2009
  • 页码 123 p.
  • 总页数 123
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号