کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
425104 685687 2013 12 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
G-Hadoop: MapReduce across distributed data centers for data-intensive computing
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر نظریه محاسباتی و ریاضیات
پیش نمایش صفحه اول مقاله
G-Hadoop: MapReduce across distributed data centers for data-intensive computing
چکیده انگلیسی

Recently, the computational requirements for large-scale data-intensive analysis of scientific data have grown significantly. In High Energy Physics (HEP) for example, the Large Hadron Collider (LHC) produced 13 petabytes of data in 2010. This huge amount of data is processed on more than 140 computing centers distributed across 34 countries. The MapReduce paradigm has emerged as a highly successful programming model for large-scale data-intensive computing applications. However, current MapReduce implementations are developed to operate on single cluster environments and cannot be leveraged for large-scale distributed data processing across multiple clusters. On the other hand, workflow systems are used for distributed data processing across data centers. It has been reported that the workflow paradigm has some limitations for distributed data processing, such as reliability and efficiency. In this paper, we present the design and implementation of G-Hadoop, a MapReduce framework that aims to enable large-scale distributed computing across multiple clusters.


► The G-Hadoop system enables massive data processing across multiple clusters and datacenters.
► The G-Hadoop system adopts existing multi-datacenter infrastructures.
► The G-Hadoop system provides the MapReduce programming model.
► G-Hadoop keeps the Hadoop API unchanged and users can re-use existing codes.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: Future Generation Computer Systems - Volume 29, Issue 3, March 2013, Pages 739–750
نویسندگان
, , , , , , ,