首页> 外文会议>Advances in Artificial Life >Evolution of Reinforcement Learning in Uncertain Environments: Emergence of Risk-Aversion and Matching
【24h】

Evolution of Reinforcement Learning in Uncertain Environments: Emergence of Risk-Aversion and Matching

机译:不确定环境下强化学习的演变:风险规避和匹配的出现

获取原文

摘要

Reinforcement learning (RL) is a fundamental process by which organisms learn to achieve a goal from interactions with the environment. Using Artificial Life techniques we derive (near-)optimal neu-ronal learning rules in a simple neural network model of decision-making in simulated bumblebees foraging for nectar. The resulting networks exhibit efficient RL, allowing the bees to respond rapidly to changes in reward contingencies. The evolved synaptic plasticity dynamics give rise to varying exploration/exploitation levels from which emerge the well-documented foraging strategies of risk aversion and probability matching. These are shown to be a direct result of optimal RL, providing a biologically founded, parsimonious and novel explanation for these behaviors. Our results are corroborated by a rigorous mathematical analysis and by experiments in mobile robots.
机译:强化学习(RL)是生物学习与环境互动来实现目标的基本过程。使用人工生命技术,我们在模拟大黄蜂觅食花蜜的简单神经网络决策模型中,得出(近)最优的神经元学习规则。由此产生的网络具有有效的RL,从而使蜜蜂能够对奖励突发事件的变化做出快速响应。进化的突触可塑性动力学导致变化的勘探/开发水平,由此产生了有据可查的风险规避和概率匹配的觅食策略。这些被证明是最佳RL的直接结果,为这些行为提供了生物学基础,简约和新颖的解释。严格的数学分析和移动机器人的实验证实了我们的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号