首页> 美国卫生研究院文献>Frontiers in Neurorobotics >DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
【2h】

DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information

机译:DeepDynamicHand:用于在利用时间信息的视频源中标注手操纵策略的深度神经体系结构

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Humans are capable of complex manipulation interactions with the environment, relying on the intrinsic adaptability and compliance of their hands. Recently, soft robotic manipulation has attempted to reproduce such an extraordinary behavior, through the design of deformable yet robust end-effectors. To this goal, the investigation of human behavior has become crucial to correctly inform technological developments of robotic hands that can successfully exploit environmental constraint as humans actually do. Among the different tools robotics can leverage on to achieve this objective, deep learning has emerged as a promising approach for the study and then the implementation of neuro-scientific observations on the artificial side. However, current approaches tend to neglect the dynamic nature of hand pose recognition problems, limiting the effectiveness of these techniques in identifying sequences of manipulation primitives underpinning action generation, e.g., during purposeful interaction with the environment. In this work, we propose a vision-based supervised Hand Pose Recognition method which, for the first time, takes into account temporal information to identify meaningful sequences of actions in grasping and manipulation tasks. More specifically, we apply Deep Neural Networks to automatically learn features from hand posture images that consist of frames extracted from grasping and manipulation task videos with objects and external environmental constraints. For training purposes, videos are divided into intervals, each associated to a specific action by a human supervisor. The proposed algorithm combines a Convolutional Neural Network to detect the hand within each video frame and a Recurrent Neural Network to predict the hand action in the current frame, while taking into consideration the history of actions performed in the previous frames. Experimental validation has been performed on two datasets of dynamic hand-centric strategies, where subjects regularly interact with objects and environment. Proposed architecture achieved a very good classification accuracy on both datasets, reaching performance up to 94%, and outperforming state of the art techniques. The outcomes of this study can be successfully applied to robotics, e.g., for planning and control of soft anthropomorphic manipulators.
机译:人类能够依靠其内在的适应性和顺应性来与环境进行复杂的操纵交互。近来,软机器人操纵试图通过设计可变形但坚固的末端执行器来重现这种非凡的行为。为了实现这一目标,对人类行为的研究对于正确地告知机器人手的技术发展已变得至关重要,该技术开发可以像人类实际一样成功地利用环境约束。在机器人可以利用的不同工具中,深度学习已成为一种有前途的研究方法,然后在人工方面进行神经科学观察。然而,当前的方法趋于忽略手势识别问题的动态性质,从而限制了这些技术在识别支撑动作产生的操作原语序列时(例如在与环境的有目的相互作用期间)的有效性。在这项工作中,我们提出了一种基于视觉的监督式手部姿势识别方法,该方法首次考虑了时间信息,以识别在抓取和操纵任务中有意义的动作序列。更具体地说,我们应用深度神经网络从手势图像中自动学习特征,这些手势图像是从具有物体和外部环境约束的抓握和操纵任务视频中提取的帧组成的。出于培训目的,将视频分为多个间隔,每个间隔都与人工监督员的特定操作相关联。所提出的算法结合了卷积神经网络来检测每个视频帧中的手和递归神经网络来预测当前帧中的手动作,同时考虑了先前帧中执行的动作的历史记录。实验验证已在两个以手为中心的动态策略的数据集上进行,受试者在其中定期与物体和环境互动。拟议的体系结构在两个数据集上都实现了非常好的分类精度,达到了高达94%的性能,并且性能优于最新技术。这项研究的成果可以成功地应用于机器人技术,例如用于规划和控制软拟人操纵器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号