...
首页> 外文期刊>Neural Networks: The Official Journal of the International Neural Network Society >Two forms of immediate reward reinforcement learning for exploratory data analysis.
【24h】

Two forms of immediate reward reinforcement learning for exploratory data analysis.

机译:两种形式的即时奖励强化学习用于探索性数据分析。

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

We review two forms of immediate reward reinforcement learning: in the first of these, the learner is a stochastic node while in the second the individual unit is deterministic but has stochastic synapses. We illustrate the first method on the problem of Independent Component Analysis. Four learning rules have been developed from the second perspective and we investigate the use of these learning rules to perform linear projection techniques such as principal component analysis, exploratory projection pursuit and canonical correlation analysis. The method is very general and simply requires a reward function which is specific to the function we require the unit to perform. We also discuss how the method can be used to learn kernel mappings and conclude by illustrating its use on a topology preserving mapping.
机译:我们回顾了两种形式的即时奖励强化学习:在第一种中,学习者是一个随机节点,而在第二种中,单个单元是确定性的但具有随机突触。我们说明了关于独立成分分析问题的第一种方法。从第二个角度已经开发了四个学习规则,我们研究了使用这些学习规则来执行线性投影技术,例如主成分分析,探索性投影追踪和规范相关分析。该方法非常通用,仅需要奖励功能,该功能特定于我们要求单位执行的功能。我们还将讨论如何将该方法用于学习内核映射,并通过说明其在拓扑保留映射上的用法来得出结论。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号