Crowdsourcing feedback for pay-as-you-go data integration

ISSN: 16130073
11Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Providing an integrated representation of data from heterogeneous data sources involves the specification of mappings that transform the data into a consistent logical schema. With a view to supporting large-scale data integration, the specification of such mappings can be carried out automatically using algorithms and heuristics. However, automatically generated mappings typically provide partial and/or incorrect results. Users can help to improve such mappings; expert users can act on the mappings directly using data integration tools, and end users or crowds can provide feedback in a pay-as-you-go fashion on results from the mappings. Such feedback can be used to inform the selection and refinement of mappings, thus improving the quality of the integration and reducing the need for expensive and potentially scarce expert staff. In this paper, we investigate the use of crowdsourcing to obtain feedback on mapping results that inform mapping selection and refinement. The investigation involves an experiment in Amazon Mechanical Turk that obtains feedback from the crowd on the correctness of mapping results. The paper describes this experiment, considers generic issues such as reliability, and reports the results for different mappings and reliability strategies.

Cite

CITATION STYLE

APA

Osorno-Gutierrez, F., Paton, N. W., & Fernandes, A. A. A. (2013). Crowdsourcing feedback for pay-as-you-go data integration. In CEUR Workshop Proceedings (Vol. 1025, pp. 32–37). CEUR-WS.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free