首页> 外文期刊>Mathematical Problems in Engineering: Theory, Methods and Applications >Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges
【24h】

Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges

机译:基于深度学习的抽象文本摘要:方法,数据集,评估措施和挑战

获取原文
           

摘要

In recent years, the volume of textual data has rapidly increased, which has generated a valuable resource for extracting and analysing information. To retrieve useful knowledge within a reasonable time period, this information must be summarised. This paper reviews recent approaches for abstractive text summarisation using deep learning models. In addition, existing datasets for training and validating these approaches are reviewed, and their features and limitations are presented. The Gigaword dataset is commonly employed for single-sentence summary approaches, while the Cable News Network (CNN)/Daily Mail dataset is commonly employed for multisentence summary approaches. Furthermore, the measures that are utilised to evaluate the quality of summarisation are investigated, and Recall-Oriented Understudy for Gisting Evaluation 1 (ROUGE1), ROUGE2, and ROUGE-L are determined to be the most commonly applied metrics. The challenges that are encountered during the summarisation process and the solutions proposed in each approach are analysed. The analysis of the several approaches shows that recurrent neural networks with an attention mechanism and long short-term memory (LSTM) are the most prevalent techniques for abstractive text summarisation. The experimental results show that text summarisation with a pretrained encoder model achieved the highest values for ROUGE1, ROUGE2, and ROUGE-L (43.85, 20.34, and 39.9, respectively). Furthermore, it was determined that most abstractive text summarisation models faced challenges such as the unavailability of a golden token at testing time, out-of-vocabulary (OOV) words, summary sentence repetition, inaccurate sentences, and fake facts.
机译:近年来,文本数据的数量迅速增加,这已经产生了有价值的资源来提取和分析信息。要在合理的时间段内检索有用的知识,必须概括此信息。本文评论了最近使用深度学习模型的抽象文本汇总方法。此外,审查了用于培训和验证这些方法的现有数据集,并提出了它们的特征和限制。 Gigaword DataSet通常用于单句摘要方法,而电缆新闻网(CNN)/每日邮件数据集通常用于多学期摘要方法。此外,研究了用于评估总结质量的措施,并召回针对直肠评估1(Rouge1),Rouge2和Rouge-L的召回的升高为最常用的指标。分析了在汇总过程中遇到的挑战和在每种方法中提出的解决方案。对几种方法的分析表明,具有注意机制和长短期记忆(LSTM)的经常性神经网络是抽象文本汇总的最普遍的技术。实验结果表明,用普拉普雷雷达编码器模型的文本汇总实现了Rouge1,Rouge2和Rouge-L(43.85,20.34和39.9)的最高值。此外,确定最具抽象文本汇总模型面临挑战,例如在测试时间,词汇(OOV)单词,摘要句子重复,不准确的句子和假事实中的金色令牌的不可用品等挑战。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号