ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

PROCESSING IMAGE FILES USING SEQUENCE FILE IN HADOOP

Journal: International Journal of Engineering Sciences & Research Technology (IJESRT) (Vol.5, No. 10)

Publication Date:

Authors : ; ; ;

Page : 521-528

Keywords : MapReduce; distributed data processing; Hadoop; sequence file;

Source : Downloadexternal Find it from : Google Scholarexternal

Abstract

This paper presents MapReduce as a distributed data processing model utilizing open source Hadoop framework for work huge volume of data. The expansive volume of data in the advanced world, especially multimedia data, makes new requirement for processing and storage. As an open source distributed compu tational framework, Hadoop takes into consideration processing a lot of images on an unbounded arrangement of computing nodes by giving fundamental foundations. We have lots and lots of small images files and need to remove duplicate files from the availab le data. As most binary formats ? particularly those that are compressed or encrypted ? cannot be split and must be read as a single linear stream of data. Using such files as input to a MapReduce job means that a single mapper will be used to process the enti re file, causing a potentially large performance hit. The paper proposes splitable format such as SequenceFile and uses MD5 algorithm to improve the p erformance of image processing.

Last modified: 2016-10-15 21:36:40