ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

Design and Implementation of a Two Level Scheduler for HADOOP Data Grids

Journal: International Journal of Advanced Networking and Applications (Vol.1, No. 05)

Publication Date:

Authors : ; ;

Page : 295-300

Keywords : Hadoop; scheduling; fair share scheduler; backfilling;

Source : Downloadexternal Find it from : Google Scholarexternal

Abstract

Hadoop is a large scale distributed processing infrastructure designed to handle data intensive applications. In a commercial large scale cluster framework, a scheduler distributes user jobs evenly among the cluster resources. The proposed work enhances Hadoop’s fair scheduler that queues the jobs for execution in a fine grained manner using task scheduling. In contrast, the proposed approach allows backfilling of jobs submitted to the scheduler. Thus job level and task level scheduling is enabled by this approach. The jobs are fairly scheduled with fairness among users, pools and priority. The outcome of the proposed work is that short narrow jobs will be executed in the slot if sufficient resource is not available for larger jobs. Thus shorter jobs get executed faster by the scheduler when compared to the existing fair scheduling policy that schedules tasks based on their fairness of remaining execution time. This approach prevents the starvation of smaller jobs if sufficient resources are available.

Last modified: 2015-12-05 20:19:44