首页> 外文会议>9th International conference on language resources and evaluation >Towards Linked Hypernyms Dataset 2.0: Complementing DBpedia with Hypernym Discovery and Statistical Type Inference
【24h】

Towards Linked Hypernyms Dataset 2.0: Complementing DBpedia with Hypernym Discovery and Statistical Type Inference

机译:迈向链接的Hypernyms数据集2.0:用Hypernym发现和统计类型推断补充DBpedia

获取原文

摘要

This paper presents a statistical type inference algorithm for ontology alignment, which assigns DBpedia entities with a new type (class). To infer types for a specific entity, the algorithm first identifies types that co-occur with the type the entity already has, and subsequently prunes the set of candidates for the most confident one. The algorithm has one parameter for balancing specificity/reliability of the resulting type selection. The proposed algorithm is used to complement the types in the LHD dataset, which is RDF knowledge base populated by identifying hypemyms from the free text of Wikipedia articles. The majority of types assigned to entities in LHD 1.0 are DBpedia resources. Through the statistical type inference, the number of entities with a type from DBpedia Ontology is increased significantly: by 750 thousand entities for the English dataset, 200.000 for Dutch and 440.000 for German. The accuracy of the inferred types is at 0.65 for English (as compared to 0.86 for LHD 1.0 types). A byproduct of the mapping process is a set of 11.000 mappings from DBpedia resources to DBpedia Ontology classes with associated confidence values. The number of the resulting mappings is an order of magnitude larger than what can be achieved with standard ontology alignment algorithms (Falcon, LogMapLt and YAM++), which do not utilize the type co-occurrence information. The presented algorithm is not restricted to the LHD dataset, it can be used to address generic type inference problems in presence of class membership information for a large number of instances.
机译:本文提出了一种用于本体对齐的统计类型推断算法,该算法为DBpedia实体分配了新的类型(类)。为了推断特定实体的类型,该算法首先识别与该实体已具有的类型同时出现的类型,然后修剪最有信心的候选对象集。该算法具有一个参数,用于平衡所得类型选择的特异性/可靠性。所提出的算法用于补充LHD数据集中的类型,该数据是RDF知识库,可通过从Wikipedia文章的自由文本中识别连字符来填充。在LHD 1.0中分配给实体的大多数类型是DBpedia资源。通过统计类型推断,具有DBpedia本体类型的实体数量显着增加:英语数据集增加了750万个实体,荷兰数据集增加了200.000个实体,德国数据集增加了440.000个实体。对于英语,推断类型的准确性为0.65(而对于LHD 1.0类型,则为0.86)。映射过程的副产品是从DBpedia资源到具有相关置信度值的DBpedia Ontology类的一组11.000映射。与不使用类型共现信息的标准本体对齐算法(Falcon,LogMapLt和YAM ++)相比,生成的映射数要大一个数量级。所提出的算法不限于LHD数据集,它可用于解决存在大量实例的类成员信息时的泛型类型推断问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号