首页> 外文会议>International Conference on Image, Video Processing and Artificial Intelligence >IterationNet: Accelerating model incremental update for small datasets based on knowledge distillation
【24h】

IterationNet: Accelerating model incremental update for small datasets based on knowledge distillation

机译:iterationNet:基于知识蒸馏的小型数据集加速模型增量更新

获取原文

摘要

Model iteration with new data is important to improve generalization of the model. In general, there are two methods to deal with model incremental update: (a) retraining the model with merging all data together and (b) training a separate model with the new data based on transfer learning. However, the above methods are either time-consuming or suffering from over-fitting problems when the sample size of new data is small. To address this practical issue, we propose a new iteration model, the IterationNet, which can learn features of new data while maintain the performance on the old data. It is a new model iteration method based on knowledge distillation which adds consistency network and truncate LI regularization. In classifying fake avatar images of Weibo users, IterationNet extremely decreased training time from 8 hours to 5 minutes while the accuracy rate is only reduced from 96% to 91% comparing to training with merged data. Compared with transfer learning. IterationNet showed increased accuracy rate by 21 percent with similar training time.
机译:使用新数据进行模型迭代对于改善模型的泛化非常重要。通常,有两种方法可以处理模型增量更新:(a)用基于转移学习的新数据来培训单独模型的所有数据并基于传输学习的新数据进行培训的模型。然而,当新数据的样本大小很小时,上述方法是耗时或遭受过度拟合的问题。为了解决这个实际问题,我们提出了一个新的迭代模型,迭代型号,它可以学习新数据的功能,同时保持旧数据的性能。这是一种基于知识蒸馏的新型模型迭代方法,该方法添加了一致性网络和截断Li正则化。在分类Weibo用户的虚假头像图像中,迭代网络从8小时到5分钟的培训时间将极低,而准确率仅减少到与合并数据的培训相比的96%至91%。与转移学习相比。迭代网络显示出高精度率,培训时间相似。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号