【24h】

Describing Visual Scenes using Transformed Dirichlet Processes

机译:使用变换的狄利克雷过程描述视觉场景

获取原文
获取原文并翻译 | 示例

摘要

Motivated by the problem of learning to detect and recognize objects with minimal supervision, we develop a hierarchical probabilistic model for the spatial structure of visual scenes. In contrast with most existing models, our approach explicitly captures uncertainty in the number of object instances depicted in a given image. Our scene model is based on the transformed Dirichlet process (TDP), a novel extension of the hierarchical DP in which a set of stochastically transformed mixture components are shared between multiple groups of data. For visual scenes, mixture components describe the spatial structure of visual features in an object-centered coordinate frame, while transformations model the object positions in a particular image. Learning and inference in the TDP, which has many potential applications beyond computer vision, is based on an empirically effective Gibbs sampler. Applied to a dataset of partially labeled street scenes, we show that the TDP's inclusion of spatial structure improves detection performance, flexibly exploiting partially labeled training images.
机译:出于学习在最少的监督下检测和识别对象的问题,我们针对视觉场景的空间结构开发了一个分层的概率模型。与大多数现有模型相比,我们的方法明确地捕获了给定图像中描述的对象实例数量的不确定性。我们的场景模型基于变换的Dirichlet过程(TDP),它是分层DP的新颖扩展,其中在多组数据之间共享一组随机变换的混合分量。对于视觉场景,混合分量描述了以对象为中心的坐标框中的视觉特征的空间结构,而变换则对特定图像中的对象位置进行建模。 TDP中的学习和推理基于计算机上有效的Gibbs采样器,在计算机视觉之外还有许多潜在的应用。应用于部分标记的街道场景的数据集,我们表明TDP包含的空间结构可提高检测性能,可以灵活地利用部分标记的训练图像。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号