首页> 外文会议>International Conference on Computational Linguistics >Similarity or deeper understanding? Analyzing the TED-Q dataset of evoked questions
【24h】

Similarity or deeper understanding? Analyzing the TED-Q dataset of evoked questions

机译:相似或更深入的理解? 分析诱发问题的TED-Q数据集

获取原文

摘要

We take a close look at a recent dataset of TED-talks annotated with the questions they implicitly evoke, TED-Q (Westera et al., 2020). We test to what extent the relation between a discourse and the questions it evokes is merely one of similarity or association, as opposed to deeper semantic/pragmatic interpretation. We do so by turning the TED-Q dataset into a binary classification task, constructing an analogous task from explicit questions we extract from the BookCorpus (Zhu et al., 2015), and fitting a BERT-based classifier alongside models based on different notions of similarity. The BERT-based classifier, achieving close to human performance, outperforms all similarity-based models, suggesting that there is more to identifying true evoked questions than plain similarity.
机译:我们仔细查看了最近的TED谈判数据集,其中包含了他们隐含地唤起的问题TED-Q(Westers等,2020)。 我们测试了讨论的话语与问题之间的关系在多大程度上仅仅是相似性或关联之一,而不是更深层次的语义/务实的解释。 我们这样做是通过将TED-Q数据集转换为二进制分类任务,构建来自Bookcorpus(Zhu等,2015)提取的明确问题的类似任务,并根据不同的概念与模型一起拟合基于伯特的分类器 相似性。 基于BERT的分类器,实现了接近人类性能,优于基于相似性的模型,表明还有比普通相似性识别真正的诱发问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号