首页> 外文期刊>電子情報通信学会技術研究報告 >Information-Maximization Clustering: Analytic Solution and Model Selection
【24h】

Information-Maximization Clustering: Analytic Solution and Model Selection

机译:信息最大化聚类:解析解决方案和模型选择

获取原文
获取原文并翻译 | 示例
           

摘要

A recently-proposed information-maximization clustering method (Gomes et al., NIPS2010) learns a kernel logistic regression classifier in an unsupervised manner so that mutual information between feature vectors and cluster assignments is maximized. A notable advantage of this approach is that it only involves continuous optimization of a logistic model, which is substantially easier than discrete optimization of cluster assignments. However, this method still suffers from two weaknesses: (i) manual tuning of kernel parameters is necessary, and (ii) finding a good local optimal solution is not straightforward due to the strong non-convexity of logistic-regression learning. In this paper, we first show that the kernel parameters can be systematically optimized by maximizing mutual information estimates. We then propose an alternative information-maximization clustering approach using a squared-loss variant of mutual information. This novel approach allows us to obtain clustering solutions analyti­cally in a computationally very efficient way. Through experiments, we demonstrate the usefulness of the proposed approaches.
机译:最近提出的信息最大化聚类方法(Gomes等,NIPS2010)以无监督的方式学习了内核逻辑回归分类器,从而使特征向量与聚类分配之间的互信息最大化。这种方法的显着优势是它仅涉及逻辑模型的连续优化,这比集群分配的离散优化容易得多。但是,该方法仍然存在两个缺点:(i)手动调整内核参数是必要的;(ii)由于logistic回归学习的强非凸性,很难找到一个好的局部最优解。在本文中,我们首先显示可以通过最大化互信息估计来系统地优化内核参数。然后,我们提出了一种使用互信息平方损失变量的替代信息最大化聚类方法。这种新颖的方法使我们能够以计算上非常有效的方式分析性地获得聚类解决方案。通过实验,我们证明了所提出方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号