首页> 外文会议>IEEE International Conference on Cloud Computing >The Design and Implementation of a Scalable Deep Learning Benchmarking Platform
【24h】

The Design and Implementation of a Scalable Deep Learning Benchmarking Platform

机译:可扩展深层学习基准平台的设计与实现

获取原文
获取外文期刊封面目录资料

摘要

The current Deep Learning (DL) landscape is fast-paced and is rife with non-uniform models, hardware/software (HW/SW) stacks. Currently, there is no DL benchmarking platform to facilitate the evaluation and comparison of DL innovations, be it models, frameworks, libraries, or hardware. As a result, the current practice of evaluating the benefits of proposed DL innovations is both arduous and error-prone - stifling the adoption of the innovations. In this work, we first identify 10 design features that are desirable within a DL benchmarking platform. These features include: performing the evaluation in a consistent, reproducible, and scalable manner, being framework and hardware agnostic, supporting real-world benchmarking workloads, providing in-depth model execution inspection across the HW/SW stack levels, etc. We then propose MLModelScope, a DL benchmarking platform that realizes these 10 design objectives. MLModelScope introduces a specification to define DL model evaluations and provides a runtime to provision the evaluation workflow using the user-specified HW/SW stack. MLModelScope defines abstractions for frameworks and supports the board range of DL models and evaluation scenarios. We implement MLModelScope as an open-source project with support for all major frameworks and hardware architectures. Through MLModelScope's evaluation and automated analysis workflows, we perform a case-study analysis of 37 models across 4 systems and show how model, hardware, and framework selection affects model accuracy and performance under different benchmarking scenarios. We further demonstrate how MLModelScope's tracing capability gives a holistic view of model execution and helps pinpoint bottlenecks.
机译:当前的深度学习(DL)景观快节奏,具有非均匀型号,硬件/软件(HW / SW)堆叠。目前,没有DL基准测试平台,以促进DL创新的评估和比较,是IT模型,框架,库或硬件。因此,目前评估所提出的DL创新益处的实践是艰巨和错误的 - 令人难以置疑的 - 扼杀通过创新的采用。在这项工作中,我们首先识别在DL基准测试平台内所需的10个设计特征。这些功能包括:以一致,可重复和可扩展的方式执行评估,是框架和硬件不可知论者,支持实际基准工作负载,在HW / SW堆栈级别等中提供深入的模型执行检查等。然后,我们提出MLModelscope是一种实现这10个设计目标的DL基准测试平台。 MLModelsCope引入了一种规范,用于定义DL模型评估,并提供使用用户指定的HW / SW堆栈提供评估工作流程的运行时。 mlmodelscope定义了框架的抽象,并支持DL模型和评估方案的棋盘范围。我们以支持所有主要框架和硬件架构的支持,我们将MLModelsCope作为开源项目。通过MLModelscope的评估和自动分析工作流程,我们在4个系统中执行37个型号的案例研究分析,并展示了如何在不同的基准测试方案下影响模型,硬件和框架选择对模型准确性和性能影响。我们进一步展示了Mlmodelscope的跟踪功能如何提供模型执行的整体视图,并有助于查明瓶颈。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号