首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Elbert: Fast Albert with Confidence-Window Based Early Exit
【24h】

Elbert: Fast Albert with Confidence-Window Based Early Exit

机译:ELBERT:基于信心窗口的早期出口快速艾伯特

获取原文

摘要

Despite the great success in Natural Language Processing (NLP) area, large pre-trained language models like BERT are not well-suited for resource-constrained or real-time applications owing to the large number of parameters and slow inference speed. Recently, compressing and accelerating BERT have become important topics. By incorporating a parameter-sharing strategy, ALBERT greatly reduces the number of parameters while achieving competitive performance. Nevertheless, ALBERT still suffers from a long inference time. In this work, we propose the ELBERT, which significantly improves the average inference speed compared to ALBERT due to the proposed confidence-window based early exit mechanism, without introducing additional parameters or extra training overhead. Experimental results show that ELBERT achieves an adaptive inference speedup varying from 2× to 10× with negligible accuracy degradation compared to AL-BERT on various datasets. Besides, ELBERT achieves higher accuracy than existing early exit methods used for accelerating BERT under the same computation cost. Furthermore, to understand the principle of the early exit mechanism, we also visualize the decision-making process of it in ELBERT. Our code is publicly available online.1
机译:尽管在自然语言处理(NLP)区域取得了巨大成功,但由于大量参数和慢速推理速度,伯特的大型预训练语言模型也不适合资源受限或实时应用。最近,压缩和加速BERT已成为重要的主题。通过结合参数共享策略,Albert大大减少了参数的数量,同时实现了竞争性能。尽管如此,Albert仍然遭受了长期推理的时间。在这项工作中,我们提出了ELBERT,它由于所提出的基于置信窗的早期退出机制而显着提高了与Albert相比的平均推理速度,而不会引入额外的参数或额外训练开销。实验结果表明,ELBERT与各种数据集上的AL-BERT相比,从2倍变化,从2倍实现的自适应推理加速。此外,ELBERT比在相同的计算成本下加速硼的现有早期退出方法实现更高的精度。此外,要了解早期退出机制的原理,我们还将其视为ELBERT的决策过程。我们的代码在线公开提供。 1

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号