首页> 外文会议>AAAI Conference on Artificial Intelligence >Motion-Based Generator Model: Unsupervised Disentanglement of Appearance, Trackable and Intrackable Motions in Dynamic Patterns
【24h】

Motion-Based Generator Model: Unsupervised Disentanglement of Appearance, Trackable and Intrackable Motions in Dynamic Patterns

机译:基于运动的生成器模型:动态模式中外观、可跟踪和难处理运动的无监督解纠缠

获取原文

摘要

Dynamic patterns are characterized by complex spatial and motion patterns. Understanding dynamic patterns requires a disentangled representational model that separates the factorial components. A commonly used model for dynamic patterns is the state space model, where the state evolves over time according to a transition model and the state generates the observed image frames according to an emission model. To model the motions explicitly, it is natural for the model to be based on the motions or the displacement fields of the pixels. Thus in the emission model, we let the hidden state generate the displacement field, which warps the trackable component in the previous image frame to generate the next frame while adding a simultaneously emitted residual image to account for the change that cannot be explained by the deformation. The warping of the previous image is about the trackable part of the change of image frame, while the residual image is about the intrackable part of the image. We use a maximum likelihood algorithm to learn the model parameters that iterates between inferring latent noise vectors that drive the transition model and updating the parameters given the inferred latent vectors. Meanwhile we adopt a regularization term to penalize the norms of the residual images to encourage the model to explain the change of image frames by trackable motion. Unlike existing methods on dynamic patterns, we learn our model in unsupervised setting without ground truth displacement fields or optical flows. In addition, our model defines a notion of intrackability by the separation of warped component and residual component in each image frame. We show that our method can synthesize realistic dynamic pattern, and disentangling appearance, trackable and intrackable motions. The learned models can be useful for motion transfer, and it is natural to adopt it to define and measure intrackability of a dynamic pattern.
机译:动态模式的特点是复杂的空间和运动模式。理解动态模式需要一个分离因子成分的分离表征模型。动态模式的常用模型是状态空间模型,其中状态根据过渡模型随时间演化,并且状态根据发射模型生成观察到的图像帧。为了明确地对运动进行建模,模型基于像素的运动或位移场是很自然的。因此,在发射模型中,我们让隐藏状态生成位移场,该位移场扭曲前一帧图像中的可跟踪分量以生成下一帧,同时添加同时发射的残余图像,以解释变形无法解释的变化。前一幅图像的扭曲是关于图像帧变化的可跟踪部分,而剩余图像是关于图像的难处理部分。我们使用最大似然算法来学习模型参数,这些参数在推断驱动过渡模型的潜在噪声向量和更新给定推断潜在向量的参数之间迭代。同时,我们采用正则化项来惩罚残差图像的范数,以鼓励模型通过可跟踪运动来解释图像帧的变化。与现有的动态模式方法不同,我们在无监督的环境中学习模型,没有地面真值位移场或光流。此外,我们的模型通过分离每个图像帧中的扭曲分量和残余分量来定义难处理性的概念。我们证明,我们的方法可以合成真实的动态模式,并解开外观、可跟踪和难处理的运动。所学习的模型可以用于运动传递,采用它来定义和测量动态模式的难处理性是很自然的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号