...
首页> 外文期刊>Neural computing & applications >Object detection in video sequences by a temporal modular self-adaptive SOM
【24h】

Object detection in video sequences by a temporal modular self-adaptive SOM

机译:通过时间模块化自适应SOM在视频序列中进行目标检测

获取原文
获取原文并翻译 | 示例

摘要

A video segmentation algorithm that takes advantage of using a background subtraction (BS) model with low learning rate (LLR) or a BS model with high learning rate (HLR) depending on the video scene dynamics is presented in this paper. These BS models are based on a neural network architecture, the self-organized map (SOM), and the algorithm is termed temporal modular self-adaptive SOM, TMSA_SOM. Depending on the type of scenario, the TMSA_SOM automatically classifies and processes each video into one of four different specialized modules based on an initial sequence analysis. This approach is convenient because unlike state-of-the-art (SoA) models, our proposed model solves different situations that may occur in the video scene (severe dynamic background, initial frames with dynamic objects, static background, stationary objects, etc.) with a specialized module. Furthermore, TMSA_SOM automatically identifies whether the scene has drastically changed (e.g., stationary objects of interest become dynamic or drastic illumination changes have occurred) and automatically detects when the scene has become stable again and uses this information to update the background model in a fast way. The proposed model was validated with three different video databases: Change Detection, BMC, and Wallflower. Findings showed a very competitive performance considering metrics commonly used in the literature to compare SoA models. TMSA_SOM also achieved the best results on two perceptual metrics, Ssim and D-Score, and obtained the best performance on the global quality measure, FSD (based on F-Measure, Ssim, and D-Score), demonstrating its robustness with different and complicated non-controlled scenarios. TMSA_SOM was also compared against SoA neural network approaches obtaining the best average performance on Re, Pr, and F-Measure.
机译:本文提出了一种视频分割算法,该算法根据视频场景的动态特性,利用低学习率的背景减法(BS)模型或高学习率的BS模型(HLR)。这些BS模型基于神经网络架构,自组织映射(SOM),并且该算法称为时间模块化自适应SOM,TMSA_SOM。根据场景的类型,TMSA_SOM根据初始序列分析自动将每个视频分类并处理为四个不同的专用模块之一。这种方法很方便,因为与最新(SoA)模型不同,我们提出的模型解决了视频场景中可能发生的各种情况(严重的动态背景,带有动态对象的初始帧,静态背景,固定对象等)。 )和专门的模块。此外,TMSA_SOM自动识别场景是否发生了剧烈变化(例如,静止的关注对象变为动态或发生了剧烈的照明变化),并自动检测场景何时再次变得稳定,并使用此信息以快速方式更新背景模型。所提出的模型已通过三个不同的视频数据库进行了验证:变更检测,BMC和Wallflower。考虑到文献中常用的用于比较SoA模型的指标,结果显示出非常有竞争力的表现。 TMSA_SOM在两个感知指标Ssim和D-Score上也取得了最佳结果,并且在全球质量度量FSD(基于F-Measure,Ssim和D-Score)上获得了最佳性能,展示了其在不同和复杂的非受控场景。还比较了TMSA_SOM与SoA神经网络方法,该方法在Re,Pr和F-Measure上获得了最佳的平均性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号