首页> 外文学位 >Kernel Approximation Methods for Speech Recognition
【24h】

Kernel Approximation Methods for Speech Recognition

机译:语音识别的核逼近方法

获取原文
获取原文并翻译 | 示例

摘要

Over the past five years or so, deep learning methods have dramatically improved the state of the art performance in a variety of domains, including speech recognition, computer vision, and natural language processing. Importantly, however, they suffer from a number of drawbacks: 1. Training these models is a non-convex optimization problem, and thus it is difficult to guarantee that a trained model minimizes the desired loss function. 2. These models are difficult to interpret. In particular, it is difficult to explain, for a given model, why the computations it performs make accurate predictions.;In contrast, kernel methods are straightforward to interpret, and training them is a convex optimization problem. Unfortunately, solving these optimization problems exactly is typically prohibitively expensive, though one can use approximation methods to circumvent this problem. In this thesis, we explore to what extent kernel approximation methods can compete with deep learning, in the context of large-scale prediction tasks. Our contributions are as follows: 1. We perform the most extensive set of experiments to date using kernel approximation methods in the context of large-scale speech recognition tasks, and compare performance with deep neural networks. 2. We propose a feature selection algorithm which significantly improves the performance of the kernel models, making their performance competitive with fully-connected feedforward neural networks. 3. We perform an in-depth comparison between two leading kernel approximation strategies --- random Fourier features [Rahimi and Recht, 2007] and the Nystrom method [Williams and Seeger, 2001] --- showing that although the Nystrom method is better at approximating the kernel, it performs worse than random Fourier features when used for learning.;We believe this work opens the door for future research to continue to push the boundary of what is possible with kernel methods. This research direction will also shed light on the question of when, if ever, deep models are needed for attaining strong performance.
机译:在过去的五年左右的时间里,深度学习方法已在包括语音识别,计算机视觉和自然语言处理在内的多个领域中极大地改善了现有技术的性能。但是,重要的是,它们具有许多缺点:1.训练这些模型是非凸优化问题,因此很难保证训练后的模型将所需的损失函数最小化。 2.这些模型很难解释。特别是,对于给定的模型,很难解释其执行的计算为何做出准确的预测。相反,内核方法易于解释,而训练它们是凸优化问题。不幸的是,尽管人们可以使用近似方法来规避这一问题,但要精确地解决这些优化问题通常过于昂贵。在本文中,我们探讨了在大规模预测任务的背景下,核逼近方法在多大程度上可以与深度学习竞争。我们的贡献如下:1.在大规模语音识别任务的背景下,我们使用核逼近方法执行了迄今为止最广泛的一组实验,并将性能与深度神经网络进行了比较。 2.我们提出了一种特征选择算法,该算法可以显着提高内核模型的性能,从而使其性能与完全连接的前馈神经网络相竞争。 3.我们对两种领先的核近似策略(随机傅里叶特征[Rahimi and Recht,2007]和Nystrom方法[Williams and Seeger,2001])进行了深入的比较,表明尽管Nystrom方法更好在近似内核时,它的性能要比随机傅里叶特征差。(我们认为)这项工作为将来的研究打开了大门,以继续推动内核方法可能的发展。该研究方向还将阐明何时需要深度模型来获得强大的性能。

著录项

  • 作者

    May, Avner.;

  • 作者单位

    Columbia University.;

  • 授予单位 Columbia University.;
  • 学科 Computer science.;Artificial intelligence.
  • 学位 Ph.D.
  • 年度 2018
  • 页码 155 p.
  • 总页数 155
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号