首页> 外国专利> METHOD FOR COMPRESSING NEURAL NETWORK MODEL, DEVICE, AND COMPUTER APPARATUS

METHOD FOR COMPRESSING NEURAL NETWORK MODEL, DEVICE, AND COMPUTER APPARATUS

机译:神经网络模型,装置和计算机设备的压缩方法

摘要

A method for compressing a neural network model, a device, a computer apparatus, and a computer readable medium. The method comprises: acquiring a first trained neural network model (S202); selecting one or more layers from layers of the first neural network model as layers to be compressed (S204); sorting the layers to be compressed according to a pre-determined rule (S206); and compressing, according to a sequential order from the sorting and by means of a genetic algorithm, a portion or all of the layers to be compressed, and obtaining a second neural network model (S208), wherein the accuracy of the second neural network model based on a pre-configured training sample is not less than a pre-determined accuracy value. The method, the device, the computer apparatus, and the computer readable medium compress a trained neural network model by means of a genetic algorithm, thereby reducing a computational load and storage space of the neural network model, and providing applicability of the same to apparatuses having limited memory and computational resources without compromising accuracy or compression of the neural network model.
机译:一种用于压缩神经网络模型的方法,设备,计算机设备和计算机可读介质。该方法包括:获取第一训练神经网络模型(S202);从第一神经网络模型的层中选择一个或多个层作为要压缩的层(S204);根据预定规则对要压缩的层进行分类(S206);根据排序后的顺序,并通过遗传算法压缩要压缩的部分或全部层,得到第二神经网络模型(S208),所述第二神经网络模型的准确性基于预先配置的训练样本不小于预先确定的精度值。该方法,装置,计算机设备和计算机可读介质通过遗传算法压缩训练的神经网络模型,从而减少了神经网络模型的计算量和存储空间,并提供了其对设备的适用性。具有有限的内存和计算资源,而不会影响神经网络模型的准确性或压缩性。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号