首页> 外文会议>International Joint Conference on Artificial Intelligence >A Dual Reinforcement Learning Framework for Unsupervised Text Style Transfer
【24h】

A Dual Reinforcement Learning Framework for Unsupervised Text Style Transfer

机译:无监督文本转移的双重加固学习框架

获取原文

摘要

Unsupervised text style transfer aims to transfer the underlying style of text but keep its main content unchanged without parallel data. Most existing methods typically follow two steps: first separating the content from the original style, and then fusing the content with the desired style. However, the separation in the first step is challenging because the content and style interact in subtle ways in natural language. Therefore, in this paper, we propose a dual reinforcement learning framework to directly transfer the style of the text via a one-step mapping model, without any separation of content and style. Specifically, we consider the learning of the source-to-target and target-to-source mappings as a dual task, and two rewards are designed based on such a dual structure to reflect the style accuracy and content preservation, respectively. In this way, the two one-step mapping models can be trained via reinforcement learning, without any use of parallel data. Automatic evaluations show that our model outperforms the state-of-the-art systems by a large margin, especially with more than 8 BLEU points improvement averaged on two benchmark datasets. Human evaluations also validate the effectiveness of our model in terms of style accuracy, content preservation and fluency. Our code and data, including outputs of all baselines and our model are available at https://github.com/luofuli/DualRL.
机译:无监督的文本样式传输旨在转移底层文本风格,但不会保持其主要内容不变,而无需并行数据。大多数现有方法通常遵循两个步骤:首先将内容与原始风格分开,然后用所需的样式融合内容。然而,第一步中的分离是具有挑战性的,因为内容和风格以自然语言的微妙方式互动。因此,在本文中,我们提出了一种双重加强学习框架,可以通过一步映射模型直接传输文本的样式,而无需任何内容和样式。具体地,我们考虑将源到目标和目标到源映射的学习作为双重任务,并且基于这种双结构设计了两个奖励,以分别反映风格精度和内容保存。以这种方式,可以通过加强学习培训,这两个一步式映射模型,而不使用并行数据。自动评估表明,我们的模型通过大幅度优于最先进的系统,特别是在两个基准数据集上平均超过8个BLEU积分改进。人类评估还验证了我们模型在风格准确性,内容保存和流利程度方面的有效性。我们的代码和数据,包括所有基线的输出和我们的模型可在https://github.com/luofuli/dualrl上获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号