...
首页> 外文期刊>Neurocomputing >Learning from data streams using kernel least-mean-square with multiple kernel-sizes and adaptive step-size
【24h】

Learning from data streams using kernel least-mean-square with multiple kernel-sizes and adaptive step-size

机译:使用具有多个内核大小和自适应步长大小的内核最小均方从数据流中学习

获取原文
获取原文并翻译 | 示例
           

摘要

A learning task is sequential if its data samples become available over time; kernel adaptive filters (KAFs) are sequential learning algorithms. There are three main challenges in KAFs: (1) selection of an appropriate Mercer kernel; (2) the lack of an effective method to determine kernel-sizes in an online learning context; (3) how to tune the step-size parameter. This work introduces a framework for online prediction that addresses the latter two of these open challenges. The kernel-sizes, unlike traditional KAF formulations, are both created and updated in an online sequential way. Further, to improve convergence time, we propose an adaptive step-size strategy that minimizes the mean-square-error (MSE) using a stochastic gradient algorithm. The proposed framework has been tested on three real-world data sets; results show both faster convergence to relatively low values of MSE and better accuracy when compared with KAF-based methods, long short-term memory, and recurrent neural networks. (C) 2019 Elsevier B.V. All rights reserved.
机译:如果学习任务的数据样本随时间推移可用,则该学习任务为顺序任务;内核自适应滤波器(KAF)是顺序学习算法。 KAF面临三个主要挑战:(1)选择合适的Mercer内核; (2)缺乏在线学习环境中确定内核大小的有效方法; (3)如何调整步长参数。这项工作引入了在线预测的框架,以解决这些公开挑战中的后两个挑战。内核大小与传统的KAF公式不同,都是以在线顺序方式创建和更新的。此外,为了提高收敛时间,我们提出了一种自适应步长策略,该策略使用随机梯度算法将均方误差(MSE)降至最低。所提出的框架已在三个实际数据集上进行了测试;结果表明,与基于KAF的方法,较长的短期记忆和递归神经网络相比,MSE收敛速度更快,MSE值相对较低,准确性更高。 (C)2019 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号