On using crowdsourcing and active learning to improve classification performance

Crowdsourcing is an emergent trend for general-purpose classification problem solving. Over the past decade, this notion has been embodied by enlisting a crowd of humans to help solve problems. There are a growing number of real-world problems that take advantage of this technique, such as Wikipedia...

Full description

Saved in:
Bibliographic Details
Published in2011 11th International Conference on Intelligent Systems Design and Applications pp. 469 - 474
Main Authors Costa, J., Silva, C., Antunes, M., Ribeiro, B.
Format Conference Proceeding
LanguageEnglish
Japanese
Published IEEE 01.11.2011
Subjects
Online AccessGet full text
ISSN2164-7143
DOI10.1109/ISDA.2011.6121700

Cover

More Information
Summary:Crowdsourcing is an emergent trend for general-purpose classification problem solving. Over the past decade, this notion has been embodied by enlisting a crowd of humans to help solve problems. There are a growing number of real-world problems that take advantage of this technique, such as Wikipedia, Linux or Amazon Mechanical Turk. In this paper, we evaluate its suitability for classification, namely if it can outperform state-of-the-art models by combining it with active learning techniques. We propose two approaches based on crowdsourcing and active learning and empirically evaluate the performance of a baseline Support Vector Machine when active learning examples are chosen and made available for classification to a crowd in a web-based scenario. The proposed crowdsourcing active learning approach was tested with Jester data set, a text humour classification benchmark, resulting in promising improvements over baseline results.
ISSN:2164-7143
DOI:10.1109/ISDA.2011.6121700