On using crowdsourcing and active learning to improve classification performance
Crowdsourcing is an emergent trend for general-purpose classification problem solving. Over the past decade, this notion has been embodied by enlisting a crowd of humans to help solve problems. There are a growing number of real-world problems that take advantage of this technique, such as Wikipedia...
Saved in:
Published in | 2011 11th International Conference on Intelligent Systems Design and Applications pp. 469 - 474 |
---|---|
Main Authors | , , , |
Format | Conference Proceeding |
Language | English Japanese |
Published |
IEEE
01.11.2011
|
Subjects | |
Online Access | Get full text |
ISSN | 2164-7143 |
DOI | 10.1109/ISDA.2011.6121700 |
Cover
Summary: | Crowdsourcing is an emergent trend for general-purpose classification problem solving. Over the past decade, this notion has been embodied by enlisting a crowd of humans to help solve problems. There are a growing number of real-world problems that take advantage of this technique, such as Wikipedia, Linux or Amazon Mechanical Turk. In this paper, we evaluate its suitability for classification, namely if it can outperform state-of-the-art models by combining it with active learning techniques. We propose two approaches based on crowdsourcing and active learning and empirically evaluate the performance of a baseline Support Vector Machine when active learning examples are chosen and made available for classification to a crowd in a web-based scenario. The proposed crowdsourcing active learning approach was tested with Jester data set, a text humour classification benchmark, resulting in promising improvements over baseline results. |
---|---|
ISSN: | 2164-7143 |
DOI: | 10.1109/ISDA.2011.6121700 |