【24h】

Creating Common Ground through Multimodal Simulations

机译:通过多峰模拟创造共同点

获取原文
获取原文并翻译 | 示例

摘要

The demand for more sophisticated human-computer interactions is rapidly increasing, as users become more accustomed to conversation-like interactions with their devices. In this paper, we examine this changing landscape in the context of human-machine interaction in a shared workspace to achieve a common goal. In our prototype system, people and avatars cooperate to build blocks world structures through the interaction of language, gesture, vision, and action. This provides a platform to study computational issues involved in multimodal communication. In order to establish elements of the common ground in discourse between speakers, we have created an embodied 3D simulation, enabling both the generation and interpretation of multiple modalities, including: language, gesture, and the visualization of objects moving and agents acting in their environment. The simulation is built on the modeling language VoxML, that encodes objects with rich semantic typing and action affordances, and actions themselves as multimodal programs, enabling contextually salient inferences and decisions in the environment. We illustrate this with a walk-through of multimodal communication in a shared task.
机译:随着用户越来越习惯于与设备进行类似对话的交互,对更复杂的人机交互的需求正在迅速增加。在本文中,我们在共享工作区中的人机交互环境下研究了这种不断变化的格局,以实现一个共同的目标。在我们的原型系统中,人和化身通过语言,手势,视觉和动作的相互作用来构建块状世界结构。这提供了一个研究多模式通信中涉及的计算问题的平台。为了在说话者之间的话语中建立共同点的元素,我们创建了一个包含在内的3D模拟,可以生成和解释多种模式,包括:语言,手势以及对象在其环境中移动和代理的可视化。该模拟建立在建模语言VoxML的基础上,该语言对具有丰富语义类型和操作能力的对象进行编码,并将操作本身作为多模式程序进行编程,从而可以在环境中进行上下文相关的推理和决策。我们通过在共享任务中演练多模式通信来说明这一点。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号