ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

MapReduce: Simplified Data Processing on Large Cluster

Journal: International Journal of Research and Engineering (Vol.5, No. 5)

Publication Date:

Authors : ;

Page : 399-403

Keywords : MapReduce; Large Cluster; Data Mining; Big Data; Reducers; Partitioners;

Source : Downloadexternal Find it from : Google Scholarexternal

Abstract

MapReduce is a data processing approach, where a single machine acts as a master, assigning map/reduce tasks to all the other machines attached in the cluster. Technically, it could be considered as a programming model, which is applied in generating, implementation and generating large data sets. The key concept behind MapReduce is that the programmer is required to state the current problem in two basic functions, map and reduce. The scalability is handles within the system, rather than being handled by the concerned programmer. By applying various restrictions on the applied programming style, MapReduce performs several moderated functions such fault tolerance, locality optimization, load balancing as well as massive parallelization. Intermediate k/v pairs are generated by the Map, and then fed o the reduce workers by the use of the incorporated file system. The data received by the reduce workers is then merged using the same key, to produce multiple output file to the concerned user (Dean & Ghemawat, 2008). Additionally, the programmer is only required to master and write the codes regarding the easy to understand functionality.

Last modified: 2018-06-08 07:25:23