首页> 外文会议>AAAI Conference on Artificial Intelligence >Revisiting LSTM Networks for Semi-Supervised Text Classification via Mixed Objective Function
【24h】

Revisiting LSTM Networks for Semi-Supervised Text Classification via Mixed Objective Function

机译:通过混合目标函数重新探测半监督文本分类的LSTM网络

获取原文

摘要

In this paper, we study bidirectional LSTM network for the task of text classification using both supervised and semi-supervised approaches. Several prior works have suggested that either complex pretraining schemes using unsupervised methods such as language modeling (Dai and Le 2015; Miyato, Dai, and Goodfellow 2016) or complicated models (Johnson and Zhang 2017) are necessary to achieve a high classification accuracy. However, we develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results compared with more complex approaches. Furthermore, in addition to cross-entropy loss, by using a combination of entropy minimization, adversarial, and virtual adversarial losses for both labeled and unlabeled data, we report state-of-the-art results for text classification task on several benchmark datasets. In particular, on the ACL-IMDB sentiment analysis and AG-News topic classification datasets, our method outperforms current approaches by a substantial margin. We also show the generality of the mixed objective function by improving the performance on relation extraction task.
机译:在本文中,我们使用监督和半监督方法研究了对文本分类任务的双向LSTM网络。有几个事先有效建议使用诸如语言建模(DAI和LE 2015; Miyato,Dai和2016年)或复杂的模型(Johnson和Zhang 2017)的复杂预押计划,以实现高分类准确性。但是,我们开发了一种培训策略,甚至允许简单的Bilstm模型,当培训具有跨熵损失时,与更复杂的方法相比,实现竞争结果。此外,除了交叉熵损失之外,通过使用标记和未标记的数据的熵最小化,对抗性和虚拟对手损失的组合,我们向多个基准数据集中报告文本分类任务的最先进结果。特别是,在ACL-IMDB情绪分析和AG-NEWS主题分类数据集上,我们的方法优于大量边缘的电流方法。我们还通过提高关系提取任务的性能来展示混合目标函数的一般性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号