【24h】

Realtime Generation of Audible Textures Inspired by a Video Stream

机译:实时产生由视频流启发的可听纹理

获取原文

摘要

We showcase a model to generate a soundscape from a camera stream in real time. The approach relies on a training video with an associated meaningful audio track; a granular synthesizer generates a novel sound by randomly sampling and mixing audio data from such video, favoring times-tamps whose frame is similar to the current camera frame; the semantic similarity between frames is computed by a pre-trained neural network. The demo is interactive: a user points a mobile phone to different objects and hears how the generated sound changes.
机译:我们展示了一种模型,可以实时从相机流生成Soundscape。 该方法依赖于带有相关的有意义的音轨的训练视频; 粒状合成器通过随机采样和混合来自这种视频的音频数据来产生新的声音,优先于其框架类似于当前相机框架的时间夯实; 帧之间的语义相似度由预先训练的神经网络计算。 演示是互动:用户将手机指向不同的对象,并听到产生的声音变化的变化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号