首页> 外文会议>IEEE International Conference on Software Engineering and Service Science >Learning Sentence Embeddings Based on Weighted Contexts from Unlabelled Data
【24h】

Learning Sentence Embeddings Based on Weighted Contexts from Unlabelled Data

机译:学习句子基于来自未标记数据的加权上下文嵌入

获取原文

摘要

Supervised learning and unsupervised learning are mainstream methods to solve semantic textual similarity tasks. However, it is obvious that supervised learning needs substantial labeled data which is hard to obtain in reality. Therefore, we turn our attention to construct sentence embeddings using unlabelled data due to lack of annotated data and success of unsupervised word embeddings in multiple tasks. We present a simple but efficient unsupervised learning method of sentence embeddings inspired by attention mechanism, in which weighted contexts are added to models to train distributed sentence representations inspired by word2vec. Our method outperforms state-of-the-art unsupervised models on semantic textual similarity tasks.
机译:监督学习和无监督的学习是解决语义文本相似性任务的主流方法。然而,显而易见的是,监督学习需要大量标记的数据,这很难在现实中获得。因此,我们注意到通过缺乏多种任务中缺乏注释的数据和无监督的单词嵌入的成功,使用未标记的数据来构建句子嵌入的数据。我们提出了一种简单但高效的句子嵌入式的学习方法,受到注意机制的启发,其中加权上下文被添加到模型中,以培训由Word2VEC启发的分布式句子表示。我们的方法优于语义文本相似性任务的最先进的无监督模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号