首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Orthogonality-Promoting Distance Metric Learning: Convex Relaxation and Theoretical Analysis
【24h】

Orthogonality-Promoting Distance Metric Learning: Convex Relaxation and Theoretical Analysis

机译:正交促进距离度量学习:凸松弛和理论分析

获取原文
           

摘要

Distance metric learning (DML), which learns a distance metric from labeled "similar" and "dissimilar" data pairs, is widely utilized. Recently, several works investigate orthogonality-promoting regularization (OPR), which encourages the projection vectors in DML to be close to being orthogonal, to achieve three effects: (1) high balancedness – achieving comparable performance on both frequent and infrequent classes; (2) high compactness – using a small number of projection vectors to achieve a "good" metric; (3) good generalizability – alleviating overfitting to training data. While showing promising results, these approaches suffer three problems. First, they involve solving non-convex optimization problems where achieving the global optimal is NP-hard. Second, it lacks a theoretical understanding why OPR can lead to balancedness. Third, the current generalization error analysis of OPR is not directly on the regularizer. In this paper, we address these three issues by (1) seeking convex relaxations of the original nonconvex problems so that the global optimal is guaranteed to be achievable; (2) providing a formal analysis on OPR’s capability of promoting balancedness; (3) providing a theoretical analysis that directly reveals the relationship between OPR and generalization performance. Experiments on various datasets demonstrate that our convex methods are more effective in promoting balancedness, compactness, and generalization, and are computationally more efficient, compared with the nonconvex methods.
机译:距离度量学习(DML)从标记的“相似”和“不相似”数据对中学习距离度量,已被广泛使用。最近,有几篇作品研究了正交促进正则化(OPR),它鼓励DML中的投影矢量接近正交,以实现三个效果:(1)高平衡性–在频繁和不频繁的类别上均具有可比的性能; (2)高度紧凑-使用少量投影向量来获得“良好”度量; (3)良好的通用性–减少对训练数据的过度拟合。尽管显示出令人鼓舞的结果,但是这些方法遇到三个问题。首先,它们涉及解决非凸优化问题,而要实现全局最优是NP难的。其次,它缺乏理论上的理解,为什么OPR会导致平衡。第三,OPR当前的泛化误差分析并不直接在正则化器上。在本文中,我们通过(1)寻求原始非凸问题的凸松弛来解决这三个问题,从而确保可以实现全局最优; (2)对OPR促进平衡的能力进行正式分析; (3)提供了直接揭示OPR与泛化性能之间关系的理论分析。在各种数据集上进行的实验表明,与非凸方法相比,我们的凸方法在促进平衡性,紧致性和泛化性方面更有效,并且在计算上更高效。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号