首页> 外文会议>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >A Million Tweets Are Worth a Few Points: Tuning Transformers for Customer Service Tasks
【24h】

A Million Tweets Are Worth a Few Points: Tuning Transformers for Customer Service Tasks

机译:一百万推文值得几点:用于客户服务任务的调整变压器

获取原文

摘要

In online domain-specific customer service applications, many companies struggle to deploy advanced NLP models successfully, due to the limited availability of and noise in their datasets. While prior research demonstrated the potential of migrating large open-domain pretrained models for domain-specific tasks, the appropriate (pre)training strategies have not yet been rigorously evaluated in such social media customer service settings, especially under multilingual conditions. We address this gap by (ⅰ) collecting a multilingual social media corpus containing customer service conversations (865k tweets), (ⅱ) comparing various pipelines of pretraining and fine-tuning approaches, (ⅲ) applying them on 5 different end tasks. We show that pretraining a generic multilingual transformer model on our in-domain dataset. before finetuning on specific end tasks, consistently boosts performance, especially in non-English settings.
机译:在特定于在线域的客户服务应用程序中,由于其数据集中的可用性和噪声有限,许多公司难以成功部署高级NLP模型。 虽然现有研究表明,迁移域特定任务的大型开放式博客模型的潜力,但在这种社交媒体客户服务环境中尚未严格评估适当的(前)培训策略,特别是在多语言条件下。 我们通过(Ⅰ)通过(Ⅰ)收集包含客户服务谈话(865K推文)的多语种社交媒体语料库(Ⅱ)比较预先训练和微调方法的各种管道,(Ⅲ)将它们应用于5个不同的最终任务。 我们展示了我们在域中数据集的预先绘制了一般的多语言变压器模型。 在对特定结束任务的FineTuning之前,始终如一地提升性能,尤其是非英语设置。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号