首页> 外文期刊>IEEE transactions on neural systems and rehabilitation engineering >Training an Actor-Critic Reinforcement Learning Controller for Arm Movement Using Human-Generated Rewards
【24h】

Training an Actor-Critic Reinforcement Learning Controller for Arm Movement Using Human-Generated Rewards

机译:使用人类产生的奖励训练演员关键性强化学习控制员进行手臂运动

获取原文
获取原文并翻译 | 示例
           

摘要

Functional Electrical Stimulation (FES) employs neuroprostheses to apply electrical current to the nerves and muscles of individuals paralyzed by spinal cord injury to restore voluntary movement. Neuroprosthesis controllers calculate stimulation patterns to produce desired actions. To date, no existing controller is able to efficiently adapt its control strategy to the wide range of possible physiological arm characteristics, reaching movements, and user preferences that vary over time. Reinforcement learning (RL) is a control strategy that can incorporate human reward signals as inputs to allow human users to shape controller behavior. In this paper, ten neurologically intact human participants assigned subjective numerical rewards to train RL controllers, evaluating animations of goal-oriented reaching tasks performed using a planar musculoskeletal human arm simulation. The RL controller learning achieved using human trainers was compared with learning accomplished using human-like rewards generated by an algorithm; metrics included success at reaching the specified target; time required to reach the target; and target overshoot. Both sets of controllers learned efficiently and with minimal differences, significantly outperforming standard controllers. Reward positivity and consistency were found to be unrelated to learning success. These results suggest that human rewards can be used effectively to train RL-based FES controllers.
机译:功能性电刺激(FES)使用神经假体将电流施加到脊髓损伤瘫痪的个体的神经和肌肉,以恢复自愿运动。神经假体控制器计算刺激模式以产生所需的动作。迄今为止,尚无现有控制器能够有效地使其控制策略适应各种可能的生理手臂特性,运动范围以及随时间变化的用户偏好。强化学习(RL)是一种控制策略,可以将人类奖励信号作为输入,以使人类用户能够塑造控制器行为。在本文中,十名神经学完好的人类参与者分配了主观数值奖励来训练RL控制器,评估使用平面肌肉骨骼人手臂模拟执行的目标导向的到达任务的动画。将使用人类教练员进行的RL控制器学习与使用算法产生的类似人的奖励的学习进行了比较;指标包括达到指定目标的成功程度;达到目标所需的时间;和目标超调。两组控制器都能以最小的差异高效学习,大大优于标准控制器。发现奖励的积极性和一致性与学习成功无关。这些结果表明,人类奖励可以有效地用于训练基于RL的FES控制器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号