【24h】

Issues with Entailment-based Zero-shot Text Classification

机译:基于Entailment的零拍文本分类的问题

获取原文

摘要

The general format of natural language inference (NLI) makes it tempting to be used for zero-shot text classification by casting any target label into a sentence of hypothesis and verifying whether or not it could be entailed by the input, aiming at generic classification applicable on any specified label space. In this opinion piece, we point out a few overlooked issues that are yet to be discussed in this line of work. We observe huge variance across different classification datasets amongst standard BERT-based NLI models and surprisingly find that pre-trained BERT without any fine-tuning can yield competitive performance against BERT fine-tuned for NLI. With the concern that these models heavily rely on spurious lexical patterns for prediction, we also experiment with preliminary approaches for more robust NLI, but the results are in general negative. Our observations reveal implicit but challenging difficulties in entailment-based zero-shot text classification.
机译:自然语言推理的一般格式(NLI)使其诱人用于通过将任何目标标签投入到假设的句子中,并验证它是否可以被输入所属,旨在适用的通用分类 在任何指定的标签空间上。 在这个意见作品中,我们指出了一些忽视的问题,尚未在这方面讨论。 我们在基于标准BERT的NLI模型中观察到不同分类数据集的巨大方差,令人惊讶地发现,没有任何微调的预训练伯特可以产生针对NLI的伯特微调的竞争性能。 随着这些模型严重依赖于虚拟的预测的虚假词汇模式的担忧,我们还试验了更强大的NLI的初步方法,但结果一般是负面的。 我们的观察结果揭示了基于需要的零击文本分类的隐含但挑战性困难。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号