首页> 外文期刊>Computational intelligence and neuroscience >Text Semantic Classification of Long Discourses Based on Neural Networks with Improved Focal Loss
【24h】

Text Semantic Classification of Long Discourses Based on Neural Networks with Improved Focal Loss

机译:基于神经网络的神经网络文本语义分类,改善焦损

获取原文
           

摘要

Semantic classification of Chinese long discourses is an important and challenging task. Discourse text is high-dimensional and sparse. Furthermore, when the number of classes of dataset is large, the data distribution will be seriously imbalanced. In solving these problems, we propose a novel end-to-end model called CRAFL, which is based on the convolutional layer with attention mechanism, recurrent neural networks, and improved focal loss function. First, the residual network (ResNet) extracts phrase semantic representations from word embedding vectors and reduces the dimensionality of the input matrix. Then, the attention mechanism differentiates the focus on the output of ResNet, and the long short-term memory layer learns the features of the sequences. Lastly but most significantly, we apply an improved focal loss function to mitigate the problem of data class imbalance. Our model is compared with other state-of-the-art models on the long discourse dataset, and CRAFL model has proven be more efficient for this task.
机译:中国长语的语义分类是一个重要而挑战的任务。话语文本是高维和稀疏的。此外,当数据集的类别大量大时,数据分布将严重不平衡。在解决这些问题时,我们提出了一种名为Crafl的新型端到端模型,其基于卷积层,其具有注意机制,经常性神经网络和改进的焦损函数。首先,剩余网络(Reset)从单词嵌入向量中提取短语语义表示,并降低输入矩阵的维度。然后,注意机制将重点区分开在RESET的输出上,并且长短期存储器层学习序列的特征。最后但最重要的是,我们应用改进的焦点损失函数来减轻数据类不平衡的问题。我们的模型与在长篇文章数据集上的其他最先进的模型进行比较,并且CRAFL模型已被证明对此任务更有效。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号