首页> 外文会议>Annual conference on Neural Information Processing Systems >Compressive spectral embedding: sidestepping the SVD
【24h】

Compressive spectral embedding: sidestepping the SVD

机译:压缩频谱嵌入:避开SVD

获取原文

摘要

Spectral embedding based on the Singular Value Decomposition (SVD) is a widely used "preprocessing" step in many learning tasks, typically leading to dimensionality reduction by projecting onto a number of dominant singular vectors and rescaling the coordinate axes (by a predefined function of the singular value). However, the number of such vectors required to capture problem structure grows with problem size, and even partial SVD computation becomes a bottleneck. In this paper, we propose a low-complexity compressive spectral embedding algorithm, which employs random projections and finite order polynomial expansions to compute approximations to S VD-based embedding. For an m × n matrix with T non-zeros, its time complexity is O ((T + m + n) log(m + n)), and the embedding dimension is O(log(m + n)), both of which are independent of the number of singular vectors whose effect we wish to capture. To the best of our knowledge, this is the first work to circumvent this dependence on the number of singular vectors for general SVD-based embeddings. The key to sidestepping the SVD is the observation that, for downstream inference tasks such as clustering and classification, we are only interested in using the resulting embedding to evaluate pairwise similarity metrics derived from the l_2-norm, rather than capturing the effect of the underlying matrix on arbitrary vectors as a partial SVD tries to do. Our numerical results on network datasets demonstrate the efficacy of the proposed method, and motivate further exploration of its application to large-scale inference tasks.
机译:基于奇异值分解(SVD)的谱嵌入是许多学习任务中广泛使用的“预处理”步骤,通常通过投影到多个主要奇异矢量上并重新缩放坐标轴来实现降维(通过预定义的函数)奇异值)。但是,捕获问题结构所需的此类向量的数量随问题的大小而增长,甚至部分SVD计算也成为瓶颈。在本文中,我们提出了一种低复杂度的压缩频谱嵌入算法,该算法利用随机投影和有限阶多项式展开来计算基于S VD的嵌入的近似值。对于具有T个非零的m×n矩阵,其时间复杂度为O((T + m + n)log(m + n)),而嵌入维数为O(log(m + n)),两者均它们与我们希望捕获其效果的奇异矢量的数量无关。据我们所知,这是为基于SVD的一般嵌入规避对奇异矢量数量的依赖的第一项工作。避开SVD的关键在于观察到,对于诸如聚类和分类之类的下游推理任务,我们仅感兴趣于使用所得的嵌入来评估从l_2范数得出的成对相似性度量,而不是捕获基础层的影响。作为部分SVD尝试对任意矢量进行矩阵运算。我们在网络数据集上的数值结果证明了该方法的有效性,并激发了其在大规模推理任务中的应用的进一步探索。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号