首页> 外文期刊>Automation in construction >End-to-end vision-based detection, tracking and activity analysis of earthmoving equipment filmed at ground level
【24h】

End-to-end vision-based detection, tracking and activity analysis of earthmoving equipment filmed at ground level

机译:基于端到端的视觉检测,跟踪和活动分析在地面拍摄的地球移动设备

获取原文
获取原文并翻译 | 示例
           

摘要

This paper presents a new benchmark dataset for validating vision-based methods that automatically identifies visually distinctive working activities of excavators and dump trucks from individual frames of a video sequence. Our dataset consists of 10 videos of interacting pairs of construction equipment filmed at ground level with accompanying ground truth annotations. These annotations consist of per-equipment and per-frame equipment bounding boxes that also have associated identities and activity labels. Our videos depict an excavator interacting with 1 or more dump trucks. We also propose a deep learning-based method for detecting and tracking objects based on Convolutional Neural Networks (CNNs). The tracking trajectories are fed into a Hidden Markov Model (HMM) that automatically discovers and assigns activity labels for any observed object. Our HMM method leverages trajectories to train a Gaussian Mixture Model (GMM) with which we estimate the probability density function of each activity using Support Vector Machine (SVM) classifiers. The proposed HMM also models activity duration and the transition between activities. We show that our method can accurately distinguish between individual equipment working activities. Results show 97.43% detection Average Precision (AP) for excavators and 75.29% AP for dump trucks, as well as cross-category tracking accuracy of 81.94% and tracking precision of 87.45%. Separate experiment results show activity analysis results of 86.8% accuracy for excavators and 88.5% for dump trucks. Our results show that our method can accurately conduct activity analysis and can be fused with methods that detect motion trajectories to scale to the needs of practical applications.
机译:本文介绍了一个新的基准数据集,用于验证基于视觉的方法,该方法自动识别从视频序列的各个帧的挖掘机和转储卡车的视觉上独特的工作活动。我们的数据集由10个视频组成,互相拍摄的建筑设备对,伴随着地面真相注释。这些注释包括每个设备和每个帧设备边界框,也具有相关的身份和活动标签。我们的视频描绘了一种与1个或更多自卸卡车交互的挖掘机。我们还提出了一种基于深入的学习方法,用于基于卷积神经网络(CNNS)来检测和跟踪对象。跟踪轨迹被馈送到隐藏的马尔可夫模型(HMM)中,它会自动发现并为任何观察到的对象分配活动标签。我们的HMM方法利用轨迹训练高斯混合模型(GMM),我们使用支持向量机(SVM)分类器来估计每个活动的概率密度函数。建议的嗯也模拟了活动持续时间和活动之间的过渡。我们表明我们的方法可以准确地区分各个设备工作活动。结果显示挖掘机的97.43%检测平均精度(AP),75.29%AP用于自卸卡车,以及81.94%的交叉类别跟踪精度,跟踪精度为87.45%。单独的实验结果表明,挖掘机的挖掘机精度为86.8%,转储卡车的88.5%。我们的结果表明,我们的方法可以准确地进行活动分析,可以与检测运动轨迹的方法融合,以规模为实际应用的需要。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号