IMPLEMENTATION OF PARALLEL APRIORI ALGORITHM ON HADOOP CLUSTER
Journal: International Journal of Computer Science and Mobile Computing - IJCSMC (Vol.2, No. 4)Publication Date: 2013-04-15
Authors : A. Ezhilvathani K. Raja;
Page : 513-516
Keywords : Hadoop; MapReduce; Apriori;
Abstract
Nowadays due to rapid growth of data in organizations, large scale data processing is a focal point of information technology. To deal with this advancement in data collection and storage technologies, designing and implementing large-scale parallel algorithm for Data mining is gaining more interest. In Data mining, association rule learning is a popular and well researched method for discovering interesting relations between variables in large databases. This paper aims to extract frequent patterns among set of items in the transaction databases or other repositories. Apriori algorithms have a great influence for finding frequent item sets using candidate generation. Apache Hadoop software framework is used to build the cluster. It working is based on MapReduce programming model. It is used to improve the processing of large-scale data on high performance cluster. It processes vast amount of data in parallel on large cluster of computer nodes. It provides reliable, scalable, distributed computing.
Other Latest Articles
- SECURED SEARCHING OF VALUABLE DATA IN A METRIC SPACE BASED ON SIMILARITY MEASURE
- Customer Security Issues in Cloud Computing?
- Receiver Based Geographic Multicast Routing in Ad Hoc Networks
- Coverage Analysis and Chinese Postman Algorithm for Efficient Model-Based Test Generation
- Improve Query Performance Using Effective Materialized View Selection and Maintenance: A Survey?
Last modified: 2013-05-02 17:22:24