Enhancing Data Quality using Human Computation and Crowd Sourcing
Journal: Journal of Independent Studies and Research - Computing (Vol.13, No. 1)Publication Date: 2015-06-01
Authors : Vikram Kumar Kirpalani Muhammad Ejaz Tayab;
Page : 74-80
Keywords : ;
Abstract
This paper is aimed at addressing the issues that are present in the data dumps available at DBpedia by using the concept of associations i.e. concept hierarchy to enhance the quality of those data dumps. These data dumps are extracted from Wikipedia and the issues that prevail in these data dumps is because of either the data extraction frameworks or the human error during crowd-sourcing efforts made on Wikipedia. By using Human Computation techniques and employing Crowd sourcing together with query morphing, diving deeper into this subject would become easier in a better way. One of the key issues with the datasets is the presence of multiple values in a single attribute and vice versa especially in the “Place of Birth” field of important personalities. This paper highlights the implementation process in order to solve these issues and adds a survey conducted on Crowd Sourcing to highlight its impact.
Other Latest Articles
- A Semi-supervised approach to Document Clustering with Sequence Constraints
- Standard Framework for Comparison of Graph Partitioning Techniques
- Local goverment investment expenditure in poland's viovodships: 2007-2013 financial perspective
- An Investigation on Topic Maps Based Document Classification with Unbalance Classes
- Performance Analysis of Table Driven and Event Driven Protocols for Voice and Video Services in MANET
Last modified: 2018-07-17 01:09:05