首页> 外文会议>International Conference on Pattern Recognition >Multiple Instance Learning with multiple positive and negative target concepts
【24h】

Multiple Instance Learning with multiple positive and negative target concepts

机译:具有多个积极和消极目标概念的多实例学习

获取原文

摘要

We introduce a new algorithm that maps multiple instance data using both positive and negative target concepts into a data representation suitable for standard classification. Multiple instance data are characterized by bags which are in turn characterized by a variable number of feature vectors or instances. Each bag has a known positive or negative label, but the labels of any given instances within a bag is unknown. First, we use the Fuzzy Clustering of Multiple Instance data (FCMI) algorithm to identify K+ positive target concepts, which represent points in the feature space that are close to instances from positive bags, and distant to instances from negative bags. We use a simple K-means clustering algorithm to identify K- negative target concepts that supplement the positive target concepts. Next we demonstrate how the positive and negative target concepts can be used to embed each bag, which has a variable number of instances, into a feature vector with fixed dimension. A key advantage to embedded instance space feature vectors is that standard machine learning algorithms may be used in training and testing multiple instance data. Another advantage of our embedding is that it provides a simple and intuitive interpretation of the data. We show that using our feature embedding, coupled with standard classifiers such as support vector machines or k-nearest neighbors, can outperform state-of-the-art Multiple Instance Learning classifiers on benchmark datasets.
机译:我们引入了一种新算法,该算法将使用正负目标概念的多个实例数据映射到适合标准分类的数据表示形式中。多实例数据的特征在于包,而包又具有可变数量的特征矢量或实例。每个袋子都有一个已知的正面或负面标签,但是袋子中任何给定实例的标签都是未知的。首先,我们使用多实例数据模糊聚类(FCMI)算法来识别K +个积极目标概念,这些概念表示特征空间中的点,这些点最接近正袋子的实例,而远离负袋子的实例。我们使用简单的K均值聚类算法来识别K负目标概念,以补充正目标概念。接下来,我们演示如何使用正负目标概念将具有可变实例数的每个包嵌入具有固定尺寸的特征向量中。嵌入式实例空间特征向量的关键优势在于,可以在训练和测试多实例数据中使用标准的机器学习算法。我们嵌入的另一个优点是,它提供了对数据的简单直观的解释。我们证明,使用我们的特征嵌入,再加上标准分类器(例如支持向量机或k最近邻),可以在基准数据集上胜过最新的多实例学习分类器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号