首页> 外文会议>Workshop on Speech and Language Processing for Assistive Technologies >WinkTalk: a demonstration of a multimodal speech synthesis platform linking facial expressions to expressive synthetic voices
【24h】

WinkTalk: a demonstration of a multimodal speech synthesis platform linking facial expressions to expressive synthetic voices

机译:WinkTalk:将面部表情连接到表达式的多模式语音合成平台以表达为表达合成声音

获取原文

摘要

This paper describes a demonstration of the WinkTalk system, which is a speech synthesis platform using expressive synthetic voices. With the help of a webcamera and facial expression analysis, the system allows the user to control the expressive features of the synthetic speech for a particular utterance with their facial expressions. Based on a personalised mapping between three expressive synthetic voices and the users facial expressions, the system selects a voice that matches their face at the moment of sending a message. The WinkTalk system is an early research prototype that aims to demonstrate that facial expressions can be used as a more intuitive control over expressive speech synthesis than manual selection of voice types, thereby contributing to an improved communication experience for users of speech generating devices.
机译:本文介绍了WinkTalk系统的演示,它是使用表现力合成声音的语音合成平台。在网络摄像机和面部表达分析的帮助下,系统允许用户控制与其面部表情的特定话语的合成语音的表现力特征。基于三个表达式合成声音和用户面部表情之间的个性化映射,系统在发送消息时选择与其脸部匹配的声音。 WinkTalk系统是一种早期研究原型,旨在证明面部表情可以用作比表达式语音合成的更直观的控制,而不是手动选择语音类型,从而有助于语音生成设备的用户改进的通信体验。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号