...
首页> 外文期刊>Journal of land use science >Dependency-based Siamese long short-term memory network for learning sentence representations
【24h】

Dependency-based Siamese long short-term memory network for learning sentence representations

机译:基于依赖性的暹罗长短期内存网络,用于学习句子表示

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Textual representations play an important role in the field of natural language processing (NLP). The efficiency of NLP tasks, such as text comprehension and information extraction, can be significantly improved with proper textual representations. As neural networks are gradually applied to learn the representation of words and phrases, fairly efficient models of learning short text representations have been developed, such as the continuous bag of words (CBOW) and skip-gram models, and they have been extensively employed in a variety of NLP tasks. Because of the complex structure generated by the longer text lengths, such as sentences, algorithms appropriate for learning short textual representations are not applicable for learning long textual representations. One method of learning long textual representations is the Long Short-Term Memory (LSTM) network, which is suitable for processing sequences. However, the standard LSTM does not adequately address the primary sentence structure (subject, predicate and object), which is an important factor for producing appropriate sentence representations. To resolve this issue, this paper proposes the dependency-based LSTM model (D-LSTM). The D-LSTM divides a sentence representation into two parts: a basic component and a supporting component. The D-LSTM uses a pre-trained dependency parser to obtain the primary sentence information and generate supporting components, and it also uses a standard LSTM model to generate the basic sentence components. A weight factor that can adjust the ratio of the basic and supporting components in a sentence is introduced to generate the sentence representation. Compared with the representation learned by the standard LSTM, the sentence representation learned by the DLSTM contains a greater amount of useful information. The experimental results show that the D-LSTM is superior to the standard LSTM for sentences involving compositional knowledge (SICK) data.
机译:文本表示在自然语言处理领域发挥着重要作用(NLP)。通过适当的文本表示,可以显着提高NLP任务等NLP任务的效率,例如文本理解和信息提取。由于神经网络逐渐应用于学习单词和短语的表示,已经开发了相当有效的学习短文本表示模型,例如连续的单词(Cow)和跳过克模型,它们已被广泛使用各种NLP任务。由于由较长的文本长度产生的复杂结构,例如句子,适合学习短文本表示的算法不适用于学习长篇文本表示。学习长篇文本表示的一种方法是长期短期存储器(LSTM)网络,适用于处理序列。但是,标准LSTM没有充分解决初级句子结构(主题,谓词和对象),这是产生适当的句子表示的重要因素。要解决此问题,本文提出了基于依赖性的LSTM模型(D-LSTM)。 D-LSTM将句子表示分为两部分:基本组件和支持组件。 D-LSTM使用预先训练的依赖关系解析器来获取主要句子信息并生成支持组件,并且还使用标准LSTM模型来生成基本句子组件。引入了可以调整句子中基本和支持组件的比率的权重因因子以生成句子表示。与标准LSTM学习的表示相比,DLSTM学习的句子表示包含更大的有用信息。实验结果表明,D-LSTM优于涉及组建知识(病态)数据的句子的标准LSTM。

著录项

  • 来源
    《Journal of land use science》 |2018年第3期|共14页
  • 作者单位

    Shanghai Univ Sch Comp Engn &

    Sci Shanghai Peoples R China;

    Shanghai Univ Sch Comp Engn &

    Sci Shanghai Peoples R China;

    Shanghai Univ Sch Comp Engn &

    Sci Shanghai Peoples R China;

    Zhejiang Univ Coll Comp Sci &

    Technol Hangzhou Zhejiang Peoples R China;

    Shanghai Univ Lib Shanghai Univ Shanghai Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 地球物理学;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号