首页> 美国卫生研究院文献>Ecology and Evolution >Three critical factors affecting automated image species recognition performance for camera traps
【2h】

Three critical factors affecting automated image species recognition performance for camera traps

机译:影响相机陷阱自动图像种类识别性能的三个关键因素

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Ecological camera traps are increasingly used by wildlife biologists to unobtrusively monitor an ecosystems animal population. However, manual inspection of the images produced is expensive, laborious, and time‐consuming. The success of deep learning systems using camera trap images has been previously explored in preliminary stages. These studies, however, are lacking in their practicality. They are primarily focused on extremely large datasets, often millions of images, and there is little to no focus on performance when tasked with species identification in new locations not seen during training. Our goal was to test the capabilities of deep learning systems trained on camera trap images using modestly sized training data, compare performance when considering unseen background locations, and quantify the gradient of lower bound performance to provide a guideline of data requirements in correspondence to performance expectations. We use a dataset provided by Parks Canada containing 47,279 images collected from 36 unique geographic locations across multiple environments. Images represent 55 animal species and human activity with high‐class imbalance. We trained, tested, and compared the capabilities of six deep learning computer vision networks using transfer learning and image augmentation: DenseNet201, Inception‐ResNet‐V3, InceptionV3, NASNetMobile, MobileNetV2, and Xception. We compare overall performance on “trained” locations where DenseNet201 performed best with 95.6% top‐1 accuracy showing promise for deep learning methods for smaller scale research efforts. Using trained locations, classifications with <500 images had low and highly variable recall of 0.750 ± 0.329, while classifications with over 1,000 images had a high and stable recall of 0.971 ± 0.0137. Models tasked with classifying species from untrained locations were less accurate, with DenseNet201 performing best with 68.7% top‐1 accuracy. Finally, we provide an open repository where ecologists can insert their image data to train and test custom species detection models for their desired ecological domain.
机译:野生动物生物学家越来越多地使用生态相机陷阱来监视生态系统的动物种群。但是,人工检查生成的图像既昂贵,费力又费时。先前已经在初期阶段探索了使用相机陷阱图像的深度学习系统的成功。但是,这些研究缺乏实用性。它们主要集中在非常大的数据集上,通常包含数百万张图像,并且在培训期间未见过的新位置进行物种识别时,几乎没有关注性能。我们的目标是使用适度大小的训练数据来测试在相机陷阱图像上训练的深度学习系统的功能,在考虑看不见的背景位置时比较性能,并量化下限性能的梯度,以提供与性能期望相对应的数据要求准则。我们使用加拿大公园提供的数据集,其中包含从多个环境的36个独特地理位置收集的47,279张图像。图像代表55种动物和人类活动,高度失衡。我们使用传输学习和图像增强功能训练,测试并比较了六个深度学习计算机视觉网络的功能:DenseNet201,Inception-ResNet-V3,InceptionV3,NASNetMobile,MobileNetV2和Xception。我们比较了DenseNet201在“训练有素”的位置上的整体性能,其中DenseNet201以95.6%的top-1精度表现最佳,这表明有希望将深度学习方法用于规模较小的研究工作。使用受过训练的位置,具有<500张图像的分类的召回率较低且变化很大,为0.750±0.329,而具有超过1000张图像的分类的召回率较高且稳定,达到0.971±0.0137。用于从未经训练的位置对物种进行分类的模型的准确性较低,其中DenseNet201的最佳表现为68.7%的top-1准确性。最后,我们提供了一个开放的存储库,生态学家可以在其中插入其图像数据,以针对所需的生态领域训练和测试自定义物种检测模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号