首页> 外文期刊>Computing reviews >From word to sense embeddings: a survey on vector representations of meaning
【24h】

From word to sense embeddings: a survey on vector representations of meaning

机译:从单词到意义嵌入:意义向量表示的一项调查

获取原文
获取原文并翻译 | 示例
           

摘要

One application area where deep learning methods have shown remarkable success is natural language processing (NLP). The speech recognition and language understanding of virtual agents like Alexa and Siri, automatic translation between languages, and the sentiment analysis of texts and other applications have seen significant improvements over the last decade. The underlying neural network architectures for deep learning rely on inputs being presented as vectors; this is straightforward for images, instrument readings, and many other data sources. Natural language, however, is structured as a linear sequence of words, which is not directly amenable to translation into a vector. Vector space models list words and their occurrences in documents in a large matrix, making them in principle suitable for neural networks. Initial approaches listing all words against all documents result in huge structures, leading to dimensionality reduction methods like latent semantic analysis. Another problem with the vector representation of words is the need to distinguish between different meanings of the same word.
机译:深度学习方法取得了显著成功的一个应用领域是自然语言处理(NLP)。在过去的十年中,Alexa和Siri等虚拟代理的语音识别和语言理解,语言之间的自动翻译以及文本和其他应用程序的情感分析得到了显着改善。深度学习的基础神经网络体系结构依赖于以向量表示的输入。对于图像,仪器读数和许多其他数据源而言,这很简单。然而,自然语言被构造成单词的线性序列,这不直接适合翻译成向量。向量空间模型以较大的矩阵列出单词及其在文档中的出现,原则上使其适合于神经网络。最初的方法针对所有文档列出所有单词会导致庞大的结构,从而导致诸如潜在语义分析之类的降维方法。单词的向量表示的另一个问题是需要区分同一单词的不同含义。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号