首页> 外文会议>Annual Conference on Neural Information Processing Systems >Describing Visual Scenes using Transformed Dirichlet Processes
【24h】

Describing Visual Scenes using Transformed Dirichlet Processes

机译:使用变换的Dirichlet进程描述视觉场景

获取原文

摘要

Motivated by the problem of learning to detect and recognize objects with minimal supervision, we develop a hierarchical probabilistic model for the spatial structure of visual scenes. In contrast with most existing models, our approach explicitly captures uncertainty in the number of object instances depicted in a given image. Our scene model is based on the transformed Dirichlet process (TDP), a novel extension of the hierarchical DP in which a set of stochastically transformed mixture components are shared between multiple groups of data. For visual scenes, mixture components describe the spatial structure of visual features in an object-centered coordinate frame, while transformations model the object positions in a particular image. Learning and inference in the TDP, which has many potential applications beyond computer vision, is based on an empirically effective Gibbs sampler. Applied to a dataset of partially labeled street scenes, we show that the TDP's inclusion of spatial structure improves detection performance, flexibly exploiting partially labeled training images.
机译:通过学习检测和识别具有最小监督的问题的问题,我们为视觉场景的空间结构开发了一个分层概率模型。与大多数现有模型相比,我们的方法明确地在给定图像中描绘的对象实例的数量中明确地捕获不确定性。我们的场景模型基于变换的Dirichlet过程(TDP),其分层DP的新扩展,其中一组随机转换的混合组分在多组数据之间共享。对于视觉场景,混合组件描述了以环形坐标帧中的视觉特征的空间结构,而变换模拟特定图像中的对象位置。在TDP中的学习和推理,具有超出计算机视觉的许多潜在应用的,基于经验有效的GIBBS采样器。应用于部分标记的街道场景的数据集,我们表明TDP的空间结构包括检测性能,灵活地利用部分标记的训练图像。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号