ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

IMPLEMENTATION OF PARALLEL APRIORI ALGORITHM ON HADOOP CLUSTER

Journal: International Journal of Computer Science and Mobile Computing - IJCSMC (Vol.2, No. 4)

Publication Date:

Authors : ;

Page : 513-516

Keywords : Hadoop; MapReduce; Apriori;

Source : Downloadexternal Find it from : Google Scholarexternal

Abstract

Nowadays due to rapid growth of data in organizations, large scale data processing is a focal point of information technology. To deal with this advancement in data collection and storage technologies, designing and implementing large-scale parallel algorithm for Data mining is gaining more interest. In Data mining, association rule learning is a popular and well researched method for discovering interesting relations between variables in large databases. This paper aims to extract frequent patterns among set of items in the transaction databases or other repositories. Apriori algorithms have a great influence for finding frequent item sets using candidate generation. Apache Hadoop software framework is used to build the cluster. It working is based on MapReduce programming model. It is used to improve the processing of large-scale data on high performance cluster. It processes vast amount of data in parallel on large cluster of computer nodes. It provides reliable, scalable, distributed computing.

Last modified: 2013-05-02 17:22:24