首页> 外文会议>Asian conference on intelligent information and database systems >Simultaneous Localization and Segmentation of Fish Objects Using Multi-task CNN and Dense CRF
【24h】

Simultaneous Localization and Segmentation of Fish Objects Using Multi-task CNN and Dense CRF

机译:使用多任务CNN和密集CRF的鱼目标同时定位和分割

获取原文

摘要

We propose a deep learning tool to localize fish objects in benthic underwater videos on a frame by frame basis. The deep network predicts fish object spatial coordinates and simultaneously segments the corresponding pixels of each fish object. The network follows a state of the art inception resnet v2 architecture that automatically generates informative features for object localization and mask segmentation tasks. Predicted masks are passed to dense Conditional Random Field (CRF) post-processing for contour and shape refinement. Unlike prior methods that rely on motion information to segment fish objects, our proposed method only requires RGB video frames to predict both box coordinates and object pixel masks. Independence from motion information makes our proposed model more robust to camera movements or jitters, and makes it more applicable to process underwater videos taken from unmanned water vehicles. We test the model in actual benthic underwater video frames taken from ten different sites. The proposed tool can segment fish objects despite wide camera movements, blurred underwater resolutions, and is robust to a wide variety of environments and fish species shapes.
机译:我们提出了一种深度学习工具,以逐帧为基础在底栖水下视频中定位鱼类对象。深度网络预测鱼对象的空间坐标,并同时分割每个鱼对象的相应像素。该网络遵循最先进的初始resnet v2体系结构,该体系结构自动生成用于对象定位和蒙版分割任务的信息功能。预测的遮罩将传递到密集的条件随机场(CRF)后处理中,以进行轮廓和形状细化。与依靠运动信息分割鱼对象的现有方法不同,我们提出的方法仅需要RGB视频帧来预测框坐标和对象像素蒙版。独立于运动信息使我们提出的模型对摄像机的运动或抖动更加鲁棒,并使其更适用于处理从无人水上交通工具拍摄的水下视频。我们在从十个不同地点拍摄的实际底栖水下视频帧中测试该模型。所提出的工具可以尽管照相机移动范围广,水下分辨率模糊,也可以分割鱼类对象,并且对各种环境和鱼类种类都具有较强的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号