首页> 外文会议>ACM SIGKDD international conference on knowledge discovery and data mining;KDD 10 >Compressed Fisher Linear Discriminant Analysis: Classification of Randomly Projected Data
【24h】

Compressed Fisher Linear Discriminant Analysis: Classification of Randomly Projected Data

机译:压缩Fisher线性判别分析:随机投影数据的分类

获取原文

摘要

We consider random projections in conjunction with classification, specifically the analysis of Fisher's Linear Discriminant (FLD) classifier in randomly projected data spaces.Unlike previous analyses of other classifiers in this setting, we avoid the unnatural effects that arise when one insists that all pairwise distances are approximately preserved under projection. We impose no sparsity or underlying low-dimensional structure constraints on the data; we instead take advantage of the class structure inherent in the problem. We obtain a reasonably tight upper bound on the estimated misclassification error on average over the random choice of the projection, which, in contrast to early distance preserving approaches, tightens in a natural way as the number of training examples increases. It follows that, for good generalisation of FLD, the required projection dimension grows logarithmically with the number of classes. We also show that the error contribution of a covariance misspecification is always no worse in the low-dimensional space than in the initial high-dimensional space. We contrast our findings to previous related work, and discuss our insights.
机译:我们结合分类考虑随机投影,特别是对随机投影数据空间中的Fisher线性判别(FLD)分类器进行分析。 与之前在这种情况下对其他分类器进行的分析不同,我们避免了当人们坚持认为所有成对距离都在投影下近似保留时会产生的不自然的影响。我们没有对数据施加稀疏性或底层的低维结构约束;相反,我们利用了问题固有的类结构。在投影的随机选择上,我们平均获得了估计错误分类误差的合理上限,与早期的距离保留方法相比,随着训练示例数量的增加,该方法自然地趋于严格。因此,为了更好地概括FLD,所需的投影维数随类数成对数增长。我们还表明,协方差错指定的错误贡献在低维空间中总是不比在初始高维空间中差。我们将我们的发现与以前的相关工作进行对比,并讨论我们的见解。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号