首页> 外文会议>International Conference on Big Data and Artificial Intelligence >Multi-Task Fine-Tuning on BERT Using Spelling Errors Correction for Chinese Text Classification Robustness
【24h】

Multi-Task Fine-Tuning on BERT Using Spelling Errors Correction for Chinese Text Classification Robustness

机译:用拼写错误纠正校正中文文本分类鲁棒性的多任务微调

获取原文

摘要

Spelling errors are common in our daily life and the industrial application, caused by automatic speech recognition, optional character recognition and human writing. Because of lack of robustness, Text classification models trained on clean datasets tend to perform poorly on the datasets with spelling errors. We conduct experiments to find out the influence of spelling errors on the performance of Chinese text classification and solve the Chinese text classification task with spelling errors by multi-task fine-tuning on BERT. We use spelling errors correction task to assist the text classification task. The results on four Chinese text classification datasets show that our method can effectively improve the robustness of the classification model which decrease the influence of spelling errors and prove the effectiveness of multi-task fine-tuning on BERT.
机译:拼写错误在我们的日常生活和工业应用中常见,由自动语音识别,可选的字符识别和人类写作引起。 由于缺乏稳健性,在干净数据集上培训的文本分类模型往往会在具有拼写错误的数据集上执行不良。 我们进行实验,以找出拼写错误对中文文本分类性能的影响,解决中文文本分类任务,通过伯爵多任务微调拼写错误。 我们使用拼写错误纠正任务来帮助文本分类任务。 四个中文文本分类数据集的结果表明,我们的方法可以有效提高分类模型的稳健性,这减少了拼写错误的影响,并证明了伯特多任务微调的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号