首页> 外文期刊>Journal of Visualization and Computer Animation >Synthesizing multimodal utterances for conversational agents
【24h】

Synthesizing multimodal utterances for conversational agents

机译:为对话主体合成多峰话语

获取原文
获取原文并翻译 | 示例

摘要

Conversational agents are supposed to combine speech with non-verbal modalities for intelligible multimodal utterances. In this paper, we focus on the generation of gesture and speech from XML-based descriptions of their overt form. An incremental production model is presented that combines the synthesis of synchronized gestural, verbal, and facial behaviors with mechanisms for linking them in fluent utterances with natural co-articulation and transition effects. In particular, an efficient kinematic approach for animating hand gestures from shape specifications is presented, which provides fine adaptation to temporal constraints that are imposed by cross-modal synchrony.
机译:会话代理应该将语音与非语言模式相结合,以实现可理解的多模式话语。在本文中,我们着重于从基于XML的公开形式的描述中生成手势和语音。提出了一种增量式生产模型,该模型将同步的手势,言语和面部行为的综合与将它们的流畅发音与自然共同发音和过渡效果联系起来的机制相结合。尤其是,提出了一种有效的运动学方法,可以根据形状规范为手势设置动画,该方法可以很好地适应跨模式同步所施加的时间约束。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号