TDG4Crowd: Test Data Generation for Evaluation of Aggregation Algorithms in Crowdsourcing
Abstract
In crowdsourcing, existing efforts mainly use real datasets collected from crowdsourcing as test datasets to evaluate the effectiveness of aggregation algorithms. However, these work ignore the fact that the datasets obtained by crowdsourcing are usually sparse and imbalanced due to limited budget. As a result, applying the same aggregation algorithm on different datasets often show contradicting conclusions. For example, on the RTE dataset, Dawid and Skene model performs significantly better than Majority Voting, while on the LableMe dataset, the experiments give the opposite conclusion. It is challenging to obtain comprehensive and balanced datasets at a low cost. To our best knowledge, little effort have been made to the fair evaluation of aggregation algorithms. To fill in this gap, we propose a novel method named TDG4Crowd that can automatically generate comprehensive and balanced datasets. Using Kullback Leibler divergence and Kolmogorov–Smirnov test, the experiment results show the superior of our method compared with others. Aggregation algorithms also perform more consistently on the synthetic datasets generated using our method.
Cite
Text
Fang et al. "TDG4Crowd: Test Data Generation for Evaluation of Aggregation Algorithms in Crowdsourcing." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/333Markdown
[Fang et al. "TDG4Crowd: Test Data Generation for Evaluation of Aggregation Algorithms in Crowdsourcing." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/fang2023ijcai-tdg/) doi:10.24963/IJCAI.2023/333BibTeX
@inproceedings{fang2023ijcai-tdg,
title = {{TDG4Crowd: Test Data Generation for Evaluation of Aggregation Algorithms in Crowdsourcing}},
author = {Fang, Yili and Shen, Chaojie and Gu, Huamao and Han, Tao and Ding, Xinyi},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2023},
pages = {2984-2992},
doi = {10.24963/IJCAI.2023/333},
url = {https://mlanthology.org/ijcai/2023/fang2023ijcai-tdg/}
}