Scheduling MapReduce Jobs in HPC Clusters

Download Now
Provided by: Pontifical Catholic University of Rio de Janeiro
Topic: Big Data
Format: PDF
MapReduce (MR) has become a de facto standard for large-scale data analysis. Moreover, it has also attracted the attention of the HPC community due to its simplicity, efficiency and highly scalable parallel model. However, MR implementations present some issues that may complicate its execution in existing HPC clusters, especially concerning the job submission. While on MR there are no strict parameters required to submit a job, in a typical HPC cluster, users must specify the number of nodes and amount of time required to complete the job execution.
Download Now

Find By Topic