首页> 外文会议>Workshop on Automated Event Extraction of Socio-political Events from News >Analyzing ELMo and DistilBERT on Socio-political News Classification
【24h】

Analyzing ELMo and DistilBERT on Socio-political News Classification

机译:从社会政治新闻分类分析ELMo和DistilBERT

获取原文

摘要

This study evaluates the robustness of two state-of-the-art deep contextual language representations, ELMo and DistilBERT, on supervised learning of binary protest news classification (PC) and sentiment analysis (SA) of product reviews. A "cross-context" setting is enabled using test sets that are distinct from the training data. The models are fine-tuned and fed into a Feed-Forward Neural Network (FFNN) and a Bidirectional Long Short Term Memory network (BiLSTM). Multinomial Naive Bayes (MNB) and Linear Support Vector Machine (LSVM) are used as traditional baselines. The results suggest that DistilBERT can transfer generic semantic knowledge to other domains better than ELMo. DistilBERT is also 30% smaller and 83% faster than ELMo, which suggests superiority for smaller computational training budgets. When generalization is not the utmost preference and test domain is similar to the training domain, the traditional machine learning (ML) algorithms can still be considered as more economic alternatives to deep language representations.
机译:这项研究评估了两种最新的深度上下文语言表示形式ELMo和DistilBERT在二进制抗议新闻分类(PC)和产品评论的情感分析(SA)的监督学习中的鲁棒性。使用与训练数据不同的测试集可以启用“跨上下文”设置。对模型进行微调,并将其馈入前馈神经网络(FFNN)和双向长期短期记忆网络(BiLSTM)。多项式朴素贝叶斯(MNB)和线性支持向量机(LSVM)被用作传统基线。结果表明,DistilBERT可以比ELMo更好地将通用语义知识转移到其他域。 DistilBERT还比ELMo小30%,快83%,这表明较小的计算培训预算具有优势。当泛化不是最大的偏好并且测试领域与训练领域相似时,传统的机器学习(ML)算法仍然可以被视为深度语言表示的更经济的选择。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号