...
首页> 外文期刊>Digital Signal Processing >Investigation and analysis of hyper and hypo neuron pruning to selectively update neurons during unsupervised adaptation
【24h】

Investigation and analysis of hyper and hypo neuron pruning to selectively update neurons during unsupervised adaptation

机译:超高和乳房神经元修剪的调查与分析,在无监督适应期间选择性地更新神经元

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Unseen or out-of-domain data can seriously degrade the performance of a neural network model, indicating the model's failure to generalize to unseen data. Neural net pruning can not only help to reduce a model's size but can improve the model's generalization capacity as well. Pruning approaches look for low-salient neurons that are less contributive to a model's decision and hence can be removed from the model. This work investigates if pruning approaches are successful in detecting neurons that are either high-salient (mostly active or hyper) or low-salient (barely active or hypo), and whether removal of such neurons can help to improve the model's generalization capacity. Traditional blind adaptation techniques update either the whole or a subset of layers, but have never explored selectively updating individual neurons across one or more layers. Focusing on the fully connected layers of a convolutional neural network (CNN), this work shows that it may be possible to selectively adapt certain neurons (consisting of the hyper and the hypo neurons) first, followed by a full-network fine tuning. Using the task of automatic speech recognition, this work demonstrates how the removal of hyper and hypo neurons from a model can improve the model's performance on out-of-domain speech data and how selective neuron adaptation can ensure improved performance when compared to traditional blind model adaptation. (C) 2020 Elsevier Inc. All rights reserved.
机译:看不见或域名数据可以严重降低神经网络模型的性能,表明该模型未能概括到看不见的数据。神经网络修剪不仅可以有助于降低模型的尺寸,但也可以提高模型的泛化容量。修剪方法寻找对模型决定的贡献不太贡献的低凸性神经元,因此可以从模型中删除。这项工作调查了修剪方法在检测高突出(主要是活性或高)或低突出(几乎活跃或宿经)的神经元中,以及是否可以帮助改善模型的泛化能力。传统的盲适应技术更新整个层或层子集,但从未探索过一个或多个层面的选择性更新单个神经元。专注于卷积神经网络(CNN)的完全连接层,这项工作表明,可以首先选择性地调整某些神经元(由Hypo和Hypo神经元组成),然后是全网络微调。使用自动语音识别的任务,这项工作表明如何从模型中去除超高和Hypo神经元可以改善模型在域外语音数据上的性能以及与传统盲模相比,选择性神经元适应如何确保改善的性能适应。 (c)2020 Elsevier Inc.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号