首页> 外国专利> NEURAL NETWORK MODEL COMPRESSING METHOD, CORPUS TRANSLATION METHOD AND DEVICE THEREOF

NEURAL NETWORK MODEL COMPRESSING METHOD, CORPUS TRANSLATION METHOD AND DEVICE THEREOF

机译:神经网络模型压缩方法,语料库翻译方法和装置

摘要

The present disclosure relates to a neural network model compressing method, a corpus translation method and a device thereof. The neural network model compressing method comprises: acquiring a training sample set including multiple training sample pairs, each training sample pair including source data and target data corresponding to the source data; using the source data as input and the target data as verification to train an original teacher model; training an intermediate teacher model on the basis of the training sample set and the original teacher model, one or more of the intermediate teacher models forming a teacher model set; training multiple candidate student models on the basis of the training sample set, the original teacher model and the teacher model set, the multiple candidate student models forming a student model set; and according to the training result of the multiple candidate student models, selecting one candidate student model as a target student model. By introducing multiple teacher models, a multi-level guidance on the training of a student model is achieved, thus improving the quality of the student model.
机译:本公开涉及神经网络模型压缩方法,语料库翻译方法及其装置。神经网络模型压缩方法包括:获取包括多个训练样本对的训练样本集,每个训练样本对包括源数据和对应于源数据的目标数据;使用源数据作为输入和目标数据作为训练原始教师模型的验证;在培训样本集和原始教师模型的基础上培训中级教师模型,一个或多个形成教师模型集的中间教师模型;在训练样本集的基础上培训多个候选学生模型,原始教师模型和教师模型集,多个候选学生模型形成了学生模型集;并根据多个候选学生模型的培训结果,选择一个候选学生模型作为目标学生模型。通过引入多个教师模型,实现了对学生模型培训的多级指导,从而提高了学生模型的质量。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号