首页> 外文学位 >Online development of cognitive behaviors by a robot: A case study using auditory and visual sensing.
【24h】

Online development of cognitive behaviors by a robot: A case study using auditory and visual sensing.

机译:机器人在线发展认知行为:使用听觉和视觉感知的案例研究。

获取原文
获取原文并翻译 | 示例

摘要

Audition and vision are two major sensory modalities for humans to sense and understand the world. Although a significant progress has been made in automatic speech recognition and visual object recognition, the fields still face a tremendous amount of difficulties.; Motivated by the autonomous development process of humans, we are interested in building a robot that automatically develops its auditory and visual cognitive and behavioral skills through real-time interactions with the environment, which typically involves humans. We call such a robot a developmental robot. To resolve the technical challenges for a developmental robot, three basic techniques have been developed and implemented: (1) A fast incremental principal component analysis algorithm, the complementary candid incremental principal component analysis (CCIPCA) algorithm; (2) A customized version of hierarchical discriminant analysis (HDR) for long temporal contexts; (3) A developmental architecture and algorithm that integrate multimodal sensing, action-imposed learning, reinforcement learning, and communicative learning.; Based upon the above three basic techniques, we have designed and implemented a prototype robot that learns cognitive behaviors from simple to complex: (1) Grounded speech learning. The system develops its audition-driven behaviors through physical interactions with the environment. This is the first attempt we know that realizes developmental auditory learning from unsegmented and unlabeled speech streams without using priori knowledge such as handcrafted word models or task-specific heuristic. This emulates a mode in which human children learn. (2) Task transfer. The system applies what has been learned in one context to new contexts (tasks) through verbal instructions from trainers. This is the first system of this kind in an autonomous mental development (AMD) mode. (3) Semantics learning. The system acquires simple semantics through real-time multimodal inputs (vision and audition). This is the first robot with both developmental auditory and developmental visual capabilities to do online learning. It takes advantage of the spatiotemporal continuity of the physical world during learning. Among the above three learning types, the first one uses only the reflexive capability of the proposed architecture, the second one uses an additional priming capability, and the third one uses an additional multimodal integration capability.; The work reported in this thesis serves as a starting point of the on-going research on a highly integrated developmental robot.
机译:试听和视觉是人类感知和理解世界的两种主要感觉方式。尽管在自动语音识别和视觉目标识别方面已经取得了重大进展,但是这些领域仍然面临着巨大的困难。受人类自主发展过程的激励,我们有兴趣构建一种机器人,该机器人通过与环境(通常涉及人类)的实时交互来自动发展其听觉,视觉的认知和行为技能。我们称这种机器人为发展型机器人。为了解决开发型机器人的技术难题,已经开发并实施了三种基本技术:(1)快速增量主成分分析算法,互补性坦率增量主成分分析(CCIPCA)算法; (2)针对长时间环境的自定义版本的分层判别分析(HDR); (3)一种发展架构和算法,集成了多模态感测,动作学习,强化学习和交际学习。基于以上三种基本技术,我们设计并实现了一个原型机器人,该机器人可以学习从简单到复杂的认知行为:(1)地面语音学习。该系统通过与环境的物理交互来开发其试听驱动的行为。这是我们知道的第一次尝试,可以在不使用先验知识(例如手工单词模型或特定于任务的启发式)的情况下,从未分段且未标记的语音流中实现发展性听觉学习。这模拟了人类儿童学习的模式。 (2)任务转移。该系统通过培训师的口头指示将在一种情况下学习到的内容应用于新的情况(任务)。这是自主智力开发(AMD)模式下的第一个此类系统。 (3)语义学习该系统通过实时多模式输入(视觉和听觉)获得简单的语义。这是第一个具有发育听觉和发育视觉功能的机器人可以进行在线学习。它在学习过程中利用了物理世界的时空连续性。在以上三种学习类型中,第一种学习仅使用所提出体系结构的自反能力,第二种使用附加的启动能力,第三种使用附加的多峰集成能力。本文所报道的工作是对高度集成的开发型机器人进行持续研究的起点。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号