首页> 外文会议>Joint SIGHUM workshop on computational linguistics for cultural heritage, social sciences, humanities and literature >One Size Fits All? A simple LSTM for Non-literal Token- and Construction-level Classification
【24h】

One Size Fits All? A simple LSTM for Non-literal Token- and Construction-level Classification

机译:一个尺寸适合所有?非文字令牌和建筑级分类的简单LSTM

获取原文

摘要

We tackle four different tasks of non-literal language classification: token and construction level metaphor detection, classification of idiomatic use of infinitive-verb compounds, and classification of non-literal particle verbs. One of the tasks operates on the token level, while the three other tasks classify constructions such as "hot topic" or "stehen lassen" (to allow sth. to stand vs. to abandon so.). The two metaphor detection tasks are in English, while the two non-literal language detection tasks are in German. We propose a simple context-encoding LSTM model and show that it outperforms the state-of-the-art on two tasks. Additionally, we experiment with different embeddings for the token level metaphor detection task and find that 1) their performance varies according to the genre, and 2) Mikolov et al. (2013) embeddings perform best on 3 out of 4 genres, despite being one of the simplest tested models. In summary, we present a large-scale analysis of a neural model for non-literal language classification (ⅰ) at different granularities, (ⅱ) in different languages, (ⅲ) over different non-literal language phenomena.
机译:我们解决四种不同的非文字语言分类任务:令牌和施工水平隐喻检测,惯用的异常使用动词化合物的分类,以及非文字粒子动词的分类。其中一个任务在令牌水平上运行,而另外三个任务分类为“热门话题”或“斯赫伦拉森”(允许某事)进行分类(允许某事。站立与弃权。)。这两个隐喻检测任务是英文,而两个非文字语言检测任务是德语。我们提出了一个简单的上下文编码的LSTM模型,并显示它在两项任务中表现出最先进的。此外,我们试验不同嵌入的令牌水平隐喻检测任务,并找到1)它们的性能根据类型而变化,2)Mikolov等人。 (2013)嵌入式在4种类型中的3个中最佳表现,尽管是最简单的测试模型之一。总之,我们对不同粒度的非文字语言分类(Ⅰ)的神经模型进行了大规模分析,(Ⅱ)不同的语言,(Ⅲ)不同的非文字语言现象。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号