【24h】

Implicit Representations of Meaning in Neural Language Models

机译:神经语言模型中含义的隐含表示

获取原文

摘要

Does the effectiveness of neural language models derive entirely from accurate modeling of surface word co-occurrence statistics, or do these models represent and reason about the world they describe? In BART and T5 transformer language models, we identify contextual word representations that function as models of entities and situations as they evolve throughout a discourse. These neural representations have functional similarities to linguistic models of dynamic semantics: they support a linear readout of each entity's current properties and relations, and can be manipulated with predictable effects on language generation. Our results indicate that prediction in pretrained neural language models is supported, at least in part, by dynamic representations of meaning and implicit simulation of entity state, and that this behavior can be learned with only text as training data.
机译:神经语言模型的有效性是否完全来自表面词共同发生统计的准确建模,或者这些模型代表了他们描述的世界的理由? 在BART和T5变压器语言模型中,我们确定在整个话语中发展的实体和情况的模式和情况的语境字墨。 这些神经表示与动态语义的语言模型具有功能性相似:它们支持每个实体的当前属性和关系的线性读数,并且可以用关于语言生成的可预测效果来操纵。 我们的结果表明,至少部分地支持预测预测神经语言模型的预测实体状态的含义和隐式模拟的动态表示,并且该行为可以仅用文本作为训练数据学习。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号