首页> 外文期刊>International Journal of Pattern Recognition and Artificial Intelligence >A Trajectory-Based Method for Dynamic Scene Recognition
【24h】

A Trajectory-Based Method for Dynamic Scene Recognition

机译:基于轨迹的动态场景识别方法

获取原文
获取原文并翻译 | 示例

摘要

Existing methods for dynamic scene recognition mostly use global features extracted from the entire video frame or a video segment. In this paper, a trajectory-based dynamic scene recognition method is proposed. A trajectory is formed by a pixel moving across consecutive frames of a video segment. The local regions surrounding the trajectory provide useful appearance and motion information about a portion of the video segment. The proposed method works at several stages. First, dense and evenly distributed trajectories are extracted from a video segment. Then, the fully-connected-layer features are extracted from each trajectory using a pre-trained Convolutional Neural Networks (CNNs) model, forming a feature sequence. Next, these feature sequences are fed into a Long-Short-Term-Memory (LSTM) network to learn their temporal behavior. Finally, by aggregating the information of the trajectories, a global representation of the video segment can be obtained for classification purposes. The LSTM is trained using synthetic trajectory feature sequences instead of real ones. The synthetic feature sequences are generated with a series of generative adversarial networks (GANs). In addition to classification, category-specific discriminative trajectories are located in a video segment, which help reveal what portions of a video segment are more important than others. This is achieved by formulating an optimization problem to learn discriminative part detectors for all categories simultaneously. Experimental results on two benchmark dynamic scene datasets show that the proposed method is very competitive with six other methods.
机译:现有动态场景识别方法主要使用从整个视频帧或视频段中提取的全局功能。本文提出了一种基于轨迹的动态场景识别方法。轨迹由跨越视频段的连续帧移动的像素形成。围绕轨迹的局部区域提供关于视频段的一部分的有用外观和运动信息。所提出的方法在几个阶段工作。首先,从视频段中提取密集和均匀分布的轨迹。然后,使用预先训练的卷积神经网络(CNNS)模型从每个轨迹中提取完全连接层特征,形成特征序列。接下来,将这些特征序列馈入到长期内存(LSTM)网络中以学习其时间行为。最后,通过聚合轨迹的信息,可以获得用于分类目的的全局表示的视频段。使用合成轨迹特征序列而不是真实的LSTM培训。用一系列生成的对抗性网络(GAN)产生合成特征序列。除了分类外,特定于类别的鉴别轨迹位于视频段中,有助于揭示视频段的部分比其他部分更重要。这是通过制定优化问题来同时为所有类别学习辨别部分探测器来实现的。两个基准动态场景数据集的实验结果表明,该方法与其他六种方法非常竞争。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号