首页> 外文期刊>International Journal of High Performance Computing Applications >SUPPORTING SCALABLE AND DISTRIBUTED DATA SUBSETTING AND AGGREGATION IN LARGE-SCALE SEISMIC DATA ANALYSIS
【24h】

SUPPORTING SCALABLE AND DISTRIBUTED DATA SUBSETTING AND AGGREGATION IN LARGE-SCALE SEISMIC DATA ANALYSIS

机译:在大规模地震数据分析中支持可扩展的分布式数据子集和聚合

获取原文
获取原文并翻译 | 示例
           

摘要

The ability to query and process very large, terabyte-scale datasets has become a key step in many scientific and engineering applications. In this paper, we describe the application of two middleware frameworks in an integrated fashion to provide a scalable and efficient system for execution of seismic data analysis on large datasets in a distributed environment. We investigate different strategies for efficient querying of large datasets and parallel implementations of a seismic image reconstruction algorithm. Our results on a state-of-the-art mass storage system coupled with a high-end compute cluster show that our implementation is scalable and can achieve about 2.9 Gigabytes per second data processing rate - about 70% of the maximum 4.2GB/s application-level raw I/O bandwidth of the storage platform.
机译:查询和处理非常大的TB级数据集的能力已成为许多科学和工程应用程序中的关键步骤。在本文中,我们以集成的方式描述了两种中间件框架的应用,以提供可扩展且高效的系统,以在分布式环境中对大型数据集执行地震数据分析。我们研究了有效查询大型数据集的不同策略以及地震图像重建算法的并行实现。我们在最先进的大容量存储系统上结合高端计算集群的结果表明,我们的实施具有可伸缩性,可以实现每秒2.9千兆字节的数据处理速率-约为最大4.2GB / s的70%存储平台的应用程序级原始I / O带宽。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号