Survey Paper on Big Data Processing and Hadoop Components
Journal: International Journal of Science and Research (IJSR) (Vol.5, No. 7)Publication Date: 2016-07-05
Authors : Swati M. Gavali; Supriya Sarkar;
Page : 1333-1337
Keywords : Big data; Hadoop; MapReduce; RDMS;
Abstract
The term Big Data, refers to data sets whose size (volume), complexity (variability), and rate of growth (velocity) make them difficult to capture, manage, process or analyzed. To analyze this enormous amount of data Hadoop can be used. However, processing is often time-consuming. One way to decrease response time is to executing the job partially, where an approximate, early result becomes available to the user, before completion of job. The implementation of the technique will be on top of Hadoop which will help to sample HDFS blocks uniformly. We will evaluate this technique using real-world datasets and applications and we will try to demonstrate the systems performance in terms of accuracy and time. The objective of the proposed technique is to significantly improve the performance of Hadoop MapReduce for efficient Big Data processing.
Other Latest Articles
- Optimized Ranking Framework for Information Retrieval
- Study of Different Models for Ranking Error in Ranked Set Sampling
- Optimal Design Methodology of Rectangular Heat Sinks for Electronic Cooling under Natural Convective and Radiative Heat Transfer
- Tyrosine Kinase level and White Blood Cells Count in Untreated and Treated Chronic Myelogenous Leukemia Patients with BCR ABL gene
- Ultraviolet Characteristics of a Silicon Inversion Layer for Applications in Radiometry
Last modified: 2021-07-01 14:40:32