首页> 外文会议>IEEE International Conference on Image Processing >Training With Cache: Specializing Object Detectors From Live Streams Without Overfitting
【24h】

Training With Cache: Specializing Object Detectors From Live Streams Without Overfitting

机译:使用缓存进行培训:无需过多调整即可从实时流中专门化对象检测器

获取原文

摘要

Online distillation can dynamically adapt to changes of distribution in a target domain by continuously updating a smaller student model from a live video stream. This ensures the accuracy of the student model even if distribution changes occur due to a change of content. However, online distillation degrades the overall accuracy because it causes overfitting to “current” distribution not to “recent” distribution. The student model is trained on sequential incoming data, and its model parameters are overwritten with the “current” distribution. As a result, the student model forgets the “recent” distribution. To overcome this problem, we propose a new training framework using cache. Our framework temporarily stores incoming frames and teacher model outputs in a cache and trains a student model with data selected from the cache. Since our approach trains the student model with not only incoming data but also past data, it can improve the overall accuracy while adapting to changes of distribution without overfitting. To use limited cache size efficiently, we also propose a loss-aware cache algorithm that chooses training data prioritized by its loss value. Our experiments show that training with cache improves the accuracy compared with online distillation, and the loss-aware cache algorithm outperforms a cache algorithm modeled on traditional offline training.
机译:通过不断从实时视频流中更新较小的学生模型,在线提炼可以动态地适应目标域中分布的变化。即使由于内容的变化而发生分布变化,这也可以确保学生模型的准确性。但是,在线蒸馏会降低整体准确性,因为它会导致过度拟合“当前”分布而不是“最近”分布。在连续输入数据上训练学生模型,并用“当前”分布覆盖其模型参数。结果,学生模型忘记了“最近的”分布。为了克服这个问题,我们提出了一种使用缓存的新训练框架。我们的框架将传入的帧和教师模型输出临时存储在缓存中,并使用从缓存中选择的数据训练学生模型。由于我们的方法不仅使用传入数据而且使用过去数据来训练学生模型,因此它可以提高总体准确性,同时适应分布的变化而不会过度拟合。为了有效地使用有限的缓存大小,我们还提出了一种可感知丢失的缓存算法,该算法选择根据其丢失值确定优先级的训练数据。我们的实验表明,与在线精馏相比,使用缓存进行训练可以提高准确性,并且丢失感知缓存算法的性能优于在传统离线训练中建模的缓存算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号