首页> 外文会议>International Conference on Computational Linguistics >NUT-RC: Noisy User-generated Text-oriented Reading Comprehension
【24h】

NUT-RC: Noisy User-generated Text-oriented Reading Comprehension

机译:坚果RC:嘈杂的用户生成的面向文本的阅读理解

获取原文

摘要

Reading comprehension (RC) on social media such as Twitter is a critical and challenging task due to its noisy, informal, but informative nature. Most existing RC models are developed on formal datasets such as news articles and Wikipedia documents, which severely limit their performances when directly applied to the noisy and informal texts in social media. Moreover, these models only focus on a certain type of RC, extractive or generative, but ignore the integration of them. To well address these challenges, we come up with a noisy user-generated text-oriented RC model. In particular, we first introduce a set of text normalizes to transform the noisy and informal texts to the formal ones. Then, we integrate the extractive and the generative RC model by a multi-task learning mechanism and an answer selection module. Experimental results on TweetQA demonstrate that our NUT-RC model significantly outperforms the state-of-the-art social media-oriented RC models.
机译:由于其嘈杂,非正式,但信息性的性质,读取理解(如Twitter)的社交媒体上的一个关键而挑战性的任务。 大多数现有的RC模型是在新闻文章和维基百科文件等正式数据集上开发的,这在直接应用于社交媒体中的嘈杂和非正式文本时严重限制了他们的表演。 此外,这些模型仅关注某种类型的RC,提取或生成,但忽略了它们的整合。 为了解决这些挑战,我们提出了一个嘈杂的用户生成的面向文本的RC模型。 特别是,我们首先介绍一组文本规范化以将噪声和非正式文本转换为正式的文本。 然后,我们通过多任务学习机制和答案选择模块集成提取和生成RC模型。 Tweetqa的实验结果表明,我们的坚果-RC模型显着优于最先进的社交媒体型RC模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号