...
首页> 外文期刊>International journal of semantic computing >SET EXPANSION OF CONTEXTUAL SEMANTIC RELATIONS: AN ALTERNATIVE TO FULL CORPUS ANNOTATION FOR SUPERVISED CLASSIFICATION
【24h】

SET EXPANSION OF CONTEXTUAL SEMANTIC RELATIONS: AN ALTERNATIVE TO FULL CORPUS ANNOTATION FOR SUPERVISED CLASSIFICATION

机译:上下文语义关系的集合扩展:监督分类的全语料标注的替代形式

获取原文
获取原文并翻译 | 示例

摘要

Recent approaches for classification of semantic relations are based on supervised learning using large training datasets.Due to the high cost of annotating such data and to the class imbalance problem, alternatives for minimizing the effort of full corpus annotation are required. In set expansion, one of such alternatives,given a small initial training set, new relevant instances are acquired from a large corpus. However,when dealing with contextual semantic relations, which are relations that are highly dependent on the context within the sentence,set expansion is not trivial,since instances are not directly queryable and filtering requires classification under a very restricted number of training instances.This work thus proposes a bootstrapped set expansion method for contextual semantic relations.It performs a best effort extraction using the Web,and a two-stage filtering of candidate instances,the first based on syntactic patterns and the second using a feature distance-based classifier designed for the low frequency setting.The relevance of time output is measured experimentally by using the expanded set as the training data of the supervised classification task,observing an incremental improvement in performance after each bootstrapping iteration when compared to values using the unexpanded training data.
机译:语义关系分类的最新方法是基于使用大型训练数据集的监督学习,由于注释此类数据的成本高以及类不平衡问题,因此需要其他方法来最大程度地减少全语料注释的工作量。在集合扩展中,这种选择之一是在初始训练集较小的情况下,从大型语料库中获取新的相关实例。但是,当处理上下文语义关系时,由于它们不能直接查询,并且过滤需要在非常有限数量的训练实例下进行分类,因此在高度依赖于句子中上下文的关系的上下文语义关系中,集合扩展并非易事。因此,提出了一种用于上下文语义关系的自举集合扩展方法。它使用Web进行尽力而为的提取,并对候选实例进行两阶段过滤,第一阶段基于句法模式,第二阶段使用基于特征距离的分类器,用于通过使用扩展集作为监督分类任务的训练数据,实验性地测量了时间输出的相关性,与使用未扩展训练数据的值相比,每次自举迭代后观察到的性能提升都得到了改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号