首页> 外文会议>International Conference on Smart Systems and Inventive Technology >Optimization of Hadoop MapReduce Model in cloud Computing Environment
【24h】

Optimization of Hadoop MapReduce Model in cloud Computing Environment

机译:云计算环境下Hadoop MapReduce模型的优化

获取原文

摘要

In recent years data analysis has become one of the trending topic among the researchers. Moreover, Information is the new baseline of all organization, as to grow the faster and bigger. Relevant information provides the flexibility to know the like and dislike of customer and to get the relevant information requires the analysis of huge information that is stored in various format. Hadoop constitutes of two basic model i.e. Hadoop Distributed File system (HDFS) and MapReduce, Hadoop is used for processing the huge amount of data whereas MapReduce is used for data processing. Hadoop MapReduce is one of the best platform for processing the huge data in efficient manner such as processing of web logs data. In this paper, we have proposed optimized HPMR (Hadoop MapReduce) model, which maximizes the memory utilization for the task and balances the performance between the I/O system and CPUs. HPMR contains the three phase i.e. Hadoop, Map and Reduce just like any other Hadoop model, however HPMR optimizes all three phase i.e. map, shuffle and reduce. Moreover, to optimize the memory model HPMR opts for dynamic terminology and input/output optimization is done through the dual operation. Moreover, in order to evaluate the performance of our model we have performed the Word-Count application on the Wikipedia data of size 128 Mb, 256 Mb, 512 Mb, 1 GB and 2 GB. The comparative analysis shows that our model optimizes nearly 30% better than the existing one.
机译:近年来,数据分析已成为研究人员的热门话题之一。此外,信息是所有组织的新基线,以使其增长更快,更大。相关信息提供了了解客户喜欢和不喜欢的灵活性,要获得相关信息,需要分析以各种格式存储的巨大信息。 Hadoop由两个基本模型组成,即Hadoop分布式文件系统(HDFS)和MapReduce,Hadoop用于处理大量数据,而MapReduce用于数据处理。 Hadoop MapReduce是用于以高效方式处理大量数据(例如处理Web日志数据)的最佳平台之一。在本文中,我们提出了优化的HPMR(Hadoop MapReduce)模型,该模型可最大化任务的内存利用率并平衡I / O系统与CPU之间的性能。 HPMR包含三个阶段,即Hadoop,Map和Reduce,就像其他任何Hadoop模型一样,但是HPMR优化了所有三个阶段,即map,shuffle和reduce。此外,为了优化内存模型,HPMR选择了动态术语,并且通过双重操作完成了输入/输出优化。此外,为了评估我们模型的性能,我们对Wikipedia数据大小为128 Mb,256 Mb,512 Mb,1 GB和2 GB的Word-Count应用程序进行了处理。对比分析表明,我们的模型比现有模型优化了近30%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号