首页> 外文学位 >SOVEREIGN: An autonomous neural system for incrementally learning to navigate towards a rewarded goal.
【24h】

SOVEREIGN: An autonomous neural system for incrementally learning to navigate towards a rewarded goal.

机译:SOVEREIGN:一种自主神经系统,用于逐步学习以导航到奖励目标。

获取原文
获取原文并翻译 | 示例

摘要

How do reactive and planned behaviors interact in real time? How are sequences of such behaviors released at appropriate times during autonomous navigation to realize valued goals? Controllers for both animals and mobile robots, or animats, need reactive mechanisms for exploration, and learned plans to reach goal objects once an environment becomes familiar. The SOVEREIGN (Self-Organizing, Vision, Expectation, Recognition, Emotion, Intelligent, Goal-oriented Navigation) animat model embodies these capabilities, and is tested in a 3D virtual reality environment. SOVEREIGN includes several interacting subsystems which model complementary properties of cortical What and Where processing streams and which clarify similarities between mechanisms for navigation and arm movement control. As the animal explores an environment, visual inputs are processed by networks that are sensitive to visual form and motion in the What and Where streams, respectively. Position-invariant and size-invariant recognition categories are learned by real-time incremental learning in the What stream. Estimates of target position relative to the animat are computed in the Where stream, and can activate approach movements toward the target. Motion cues from animat locomotion can elicit head-orienting movements to bring a new target into view. Approach and orienting movements are alternately performed during animat navigation. Cumulative estimates of each movement are derived from interacting proprioceptive and visual cues. Movement sequences are stored within a motor working memory. Sequences of visual categories are stored in a sensory working memory. These working memories trigger learning of sensory and motor sequence categories, or plans, which together control planned movements. Predictively effective chunk combinations are selectively enhanced via reinforcement learning when the animat is rewarded. Selected planning chunks effect a gradual transition from variable reactive exploratory movements to efficient goal-oriented planned movement sequences. Volitional signals gate interactions between model subsystems and the release of overt behaviors. The model can control different motor sequences under different motivational states and learns more efficient sequences to rewarded goals as exploration proceeds.
机译:反应性和计划性行为如何实时交互?如何在自主导航期间的适当时间释放这些行为序列,以实现有价值的目标?动物和移动机器人(或动画)的控制器都需要反应性的机制进行探索,并且一旦环境变得熟悉,就学到了计划以达到目标目标。 SOVEREIGN(自组织,视觉,期望,识别,情感,智能,面向目标的导航)动画模型体现了这些功能,并在3D虚拟现实环境中进行了测试。 SOVEREIGN包括几个交互子系统,这些子系统对皮质What和Where处理流的互补特性进行建模,并阐明导航和手臂运动控制机制之间的相似性。当动物探索环境时,视觉输入由分别对What和Where流中的视觉形式和运动敏感的网络处理。位置不变和大小不变的识别类别是通过What流中的实时增量学习来学习的。相对于动画的目标位置的估计是在Where流中计算的,并且可以激活朝目标的进近运动。 animat运动产生的运动提示可以引起头部定向运动,从而使新目标可见。在动画导航期间,交替执行接近和定向运动。每个动作的累积估计值来自相互作用的本体感受和视觉提示。运动序列存储在电机工作存储器中。视觉类别的序列存储在感觉工作记忆中。这些工作记忆触发对感觉和运动序列类别或计划的学习,它们共同控制计划的运动。当奖励有生命的动物时,通过强化学习有选择地增强了预测有效的块组合。选定的计划组块实现了从可变的反应性探索性运动到有效的面向目标的计划性运动序列的逐步过渡。自愿信号控制模型子系统之间的交互作用以及公开行为的释放。该模型可以在不同的动机状态下控制不同的运动序列,并随着探索的进行,学习更有效的序列来奖励目标。

著录项

  • 作者

    Gnadt, William Patrick.;

  • 作者单位

    Boston University.;

  • 授予单位 Boston University.;
  • 学科 Biology Neuroscience.; Psychology Cognitive.; Engineering Robotics.
  • 学位 Ph.D.
  • 年度 2008
  • 页码 211 p.
  • 总页数 211
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 神经科学;心理学;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号