...
首页> 外文期刊>Biosystems Engineering >Automatic recognition of lactating sow postures by refined two-stream RGB-D faster R-CNN
【24h】

Automatic recognition of lactating sow postures by refined two-stream RGB-D faster R-CNN

机译:通过精细的双流RGB-D自动识别哺乳母猪姿势RGB-D速度R-CNN

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This paper proposes an end-to-end refined two-stream RGB-D Faster region convolutional neural network (R-CNN) algorithm, which fuses RGB-D image features in the feature extraction stage for recognising five postures of lactating sows (standing, sitting, sternal recumbency, ventral recumbency, and lateral recumbency) in scenes at a pig farm. Based on the Faster R-CNN algorithm, two CNNs were first used to extract the RGB image features and depth image features. Then, a proposed single RGB-D region proposal network was used to generate the regions of interest (ROIs) for the two types of image feature maps in RGB-D. Next, the features of the RGB-D ROIs were extracted and merged using a feature fusion layer. Finally, the fused features of the RGB-D ROIs were input into a Fast R-CNN to obtain the recognition results. A total of 12,600 pairs of RGB-D images of five postures were obtained by a Kinect v2.0 sensor and were randomly selected from the first 21 of 28 pens as the training set, and 5533 pairs were randomly selected from the remaining 7 pens as the test set. The proposed method was used to recognise the five postures of lactating sows. The recognition accuracy of the concatenation fusion method was the highest for the test set with average precisions for the five categories of lactating sow postures of 99.74%, 96.49%, 90.77%, 90.91%, and 99.45%, respectively. Compared with related methods (RGB-only method, depth-only method, RGB-D early fusion, and later fusion), our method attained the highest mean average precision. (C) 2019 IAgrE. Published by Elsevier Ltd. All rights reserved.
机译:本文提出了端到端精制的二流RGB-D更快的区域卷积神经网络(R-CNN)算法,其在特征提取阶段中熔化​​RGB-D图像特征,以识别哺乳母猪的五个姿势(站立,在养猪场的场景中坐在,胸骨闭锁,腹侧腹部和横向闭合)。基于较快的R-CNN算法,首先使用两个CNN来提取RGB图像特征和深度图像特征。然后,建议的单个RGB-D区域提议网络用于生成RGB-D中的两种图像特征映射的感兴趣区域(ROI)。接下来,使用特征融合层提取RGB-D ROI的特征并合并。最后,RGB-D ROI的融合特征被输入到快速R-CNN中以获得识别结果。通过Kinect V2.0传感器获得了5个姿势的12,600对RGB-D图像,并且从28个钢笔的前21个中随机选择,从剩余的7个钢笔随机选择5533对。测试集。该方法用于识别哺乳母猪的五个姿势。倾斜融合方法的识别准确性是测试设定的最高精度,分别为五类哺乳母猪源99.74%,96.49%,90.77%,90.91%和99.45%。与相关方法相比(仅限RGB方法,唯一的方法,RGB-D早期融合和后来的融合),我们的方法达到了最高平均平均精度。 (c)2019年IAGRE。 elsevier有限公司出版。保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号