首页> 外文会议>International Conference on Intelligent User Interfaces >Leveraging User Input and Feedback for Interactive Sound Event Detection and Annotation
【24h】

Leveraging User Input and Feedback for Interactive Sound Event Detection and Annotation

机译:利用用户输入和反馈进行交互式声音事件检测和注释

获取原文

摘要

Tagging of environment audio events is essential in many areas. However, finding sound events and labeling them within a long audio file is tedious and time-consuming. Building an automatic recognition system using modern machine learning is often not feasible because it requires a large number of human-labeled training examples and it is not reliable enough for all uses. I propose interactive sound event detection to solve the issue by combining machine search with human tagging, specifically focusing on the effectiveness of various types of user-inputs to the interactive sound searching. The types of user inputs that I will explore include binary relevance feedback, segmentation, and vocal imitation. I expect that leveraging one or combination of these user inputs would help users find audio contents of interest quickly and accurately, even in the situation where there are not enough training examples for a typical automated system.
机译:环境音频事件的标记对于许多领域至关重要。但是,在长音频文件中查找声音事件并将其标记为繁琐且耗时。使用现代机器学习构建自动识别系统通常是不可行的,因为它需要大量的人类标记的训练示例,并且对于所有用途来说是不可靠的。我提出了交互式声音事件检测来解决这些问题,通过将机器搜索与人类标记组合,专门关注各种类型用户输入的有效性来交互式声音搜索。我将浏览的用户输入类型包括二进制相关反馈,分段和声乐模仿。我希望利用这些用户输入的一个或组合可以帮助用户快速准确地找到感兴趣的音频内容,即使在没有足够的典型自动化系统的训练示例的情况下。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号