【24h】

FatMan vs. LittleBoy: Scaling Up Linear Algebraic Operations in Scale-Out Data Platforms

机译:Fatman vs. Littleboy:在扩展数据平台中缩放线性代数操作

获取原文

摘要

Linear algebraic operations such as matrix manipulations form the kernel of many machine learning and other crucial algorithms. Scaling up as well as scaling out such algorithms are highly desirable to enable efficient processing over millions of data points. To this end, we present a matrix manipulation approach to effectively scale-up each node in a scale-out data parallel platform such as Apache Spark. Specifically, we enable hardware acceleration for matrix multiplications in a distributed Spark setup without user intervention. Our approach supports both dense and sparse distributed matrices, and provides flexible control of acceleration by matrix density. We demonstrate the benefit of our approach for generalized matrix multiplication operations over large matrices with up to four billion elements. To connect the effectiveness of our approach with machine learning applications, we performed Gramian matrix computation via generalized matrix multiplications. Our experiments show that our approach achieves more than 2× performance speed-up, and up to 96.1% computation improvement, compared to a state of the art Spark MLlib for dense matrices.
机译:矩阵操作等线性代数操作形成了许多机器学习和其他关键算法的内核。缩放以及缩放此类算法非常希望能够高效地处理超过数百万数据点。为此,我们介绍了一种矩阵操纵方法,以在诸如Apache Spark之类的横向数据并行平台中有效地扩展每个节点。具体地,我们在没有用户干预的情况下,在分布式火花设置中启用矩阵乘法的硬件加速。我们的方法支持密集和稀疏的分布式矩阵,并通过矩阵密度提供灵活的加速控制。我们展示了我们对大型矩阵上的广义矩阵乘法操作的方法的好处,该矩阵高达四亿个元素。要将我们的方法与机器学习应用的效果连接,我们通过广义矩阵乘法执行克鲁米亚矩阵计算。我们的实验表明,与致密矩阵的艺术的状态相比,我们的方法达到了超过2倍的性能加速,高达96.1%的计算改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号