首页> 外文会议>Iberian conference on pattern recognition and image analysis >Uncertainty Estimation for Black-Box Classification Models: A Use Case for Sentiment Analysis
【24h】

Uncertainty Estimation for Black-Box Classification Models: A Use Case for Sentiment Analysis

机译:黑匣子分类模型的不确定性估计:情感分析的用例

获取原文

摘要

With the advent of new pre-trained word embedding models like ELMO, GPT or BERT, that leverage transfer-learning to deliver high-quality prediction systems, natural language processing (NLP) methods are reaching or even overtaking human baselines in some applications. The basic principle of these successful models is to train a model to solve a given NLP task, mainly Language Modelling, using significant volumes of data like the whole Wikipedia. The model is then fine-tuned to solve another NLP task, requiring fewer domain-specific data to achieve state-of-the-art accuracies. The method proposed in the present work assists the practitioner in evaluating the quality of the transferred classification models when applied to new data domains. In this case, we consider the original model as a black box. No matter how complex the original model may be, the method only requires access to the output layer to train a measure of the uncertainty associated with the predictions of the original model. This measure of uncertainty is a measure of how well the black-box model accommodates to the new data. Later on, we show how a rejection system can use this uncertainty to improve its accuracy, effectively enabling the practitioner to find the best trade-off between the quality of the model and the number of rejected cases.
机译:随着新的经过预训练的词嵌入模型(例如ELMO,GPT或BERT)的出现,它们利用转移学习来提供高质量的预测系统,自然语言处理(NLP)方法在某些应用中已经达到甚至超过了人类基线。这些成功模型的基本原理是使用大量数据(如整个Wikipedia)训练模型以解决给定的NLP任务,主要是语言建模。然后对该模型进行微调以解决另一项NLP任务,即需要较少的特定于域的数据来实现最新的精度。当前工作中提出的方法可帮助从业人员评估应用于新数据域的转移分类模型的质量。在这种情况下,我们将原始模型视为黑匣子。无论原始模型有多复杂,该方法仅需要访问输出层以训练与原始模型的预测相关联的不确定性的度量。这种不确定性度量是黑盒模型适应新数据的能力的度量。稍后,我们将展示拒绝系统如何利用这种不确定性来提高其准确性,从而有效地使从业者能够在模型的质量与被拒绝案例的数量之间找到最佳折衷方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号