首页> 外文期刊>Neurocomputing >Parametric nonlinear dimensionality reduction using kernel t-SNE
【24h】

Parametric nonlinear dimensionality reduction using kernel t-SNE

机译:基于核t-SNE的参数化非线性降维

获取原文
获取原文并翻译 | 示例
       

摘要

Novel non-parametric dimensionality reduction techniques such as t-distributed stochastic neighbor embedding (t-SNE) lead to a powerful and flexible visualization of high-dimensional data. One drawback of non-parametric techniques is their lack of an explicit out-of-sample extension. In this contribution, we propose an efficient extension of t-SNE to a parametric framework, kernel t-SNE, which preserves the flexibility of basic t-SNE, but enables explicit out-of-sample extensions. We test the ability of kernel t-SNE in comparison to standard t-SNE for benchmark data sets, in particular addressing the generalization ability of the mapping for novel data. In the context of large data sets, this procedure enables us to train a mapping for a fixed size subset only, mapping all data afterwards in linear time. We demonstrate that this technique yields satisfactory results also for large data sets provided missing information due to the small size of the subset is accounted for by auxiliary information such as class labels, which can be integrated into kernel t-SNE based on the Fisher information.
机译:新颖的非参数降维技术(例如t分布随机邻居嵌入(t-SNE))导致了高维数据的强大而灵活的可视化。非参数技术的一个缺点是它们缺乏明确的样本外扩展。在此贡献中,我们提出将t-SNE有效扩展到参数框架t-SNE内核,该框架保留了基本t-SNE的灵活性,但允许显式的样本外扩展。与基准数据集的标准t-SNE相比,我们测试了内核t-SNE的能力,特别是解决了新数据映射的泛化能力。在大数据集的情况下,此过程使我们能够训练仅针对固定大小子集的映射,然后在线性时间内映射所有数据。我们证明了该技术对于大数据集也能提供令人满意的结果,如果由于子集的小尺寸而丢失的信息由辅助信息(如类别标签)解决,则可以根据Fisher信息将其集成到内核t-SNE中。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号