ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

Big Data Hadoop: Aggregation Techniques

Journal: International Journal of Science and Research (IJSR) (Vol.4, No. 12)

Publication Date:

Authors : ;

Page : 432-435

Keywords : privacy preservation; security; e-healthcare systems; data mining; image feature extraction;

Source : Downloadexternal Find it from : Google Scholarexternal

Abstract

The term Big Data, refers to data sets whose size (volume), complexity (variability), and rate of growth (velocity) make them difficult to capture, manage, process or analyzed. To analyze this enormous amount of data Hadoop can be used. However, processing is often time-consuming. One way to decrease response time is to executing the job partially, where an approximate, early result becomes available to the user, before completion of job. The implementation of the technique will be on top of Hadoop which will help to sample HDFS blocks uniformly. We will evaluate this technique using real-world datasets and applications and we will try to demonstrate the systems performance in terms of accuracy and time. The objective of the proposed technique is to significantly improve the performance of Hadoop MapReduce for efficient Big Data processing.

Last modified: 2021-07-01 14:28:06