Query Design for Crowdsourced Clustering: Effect of Cognitive Overload and Contextual Bias

Abstract

Crowdsourced clustering leverages human input to group items into clusters. The design of tasks for crowdworkers, specifically the number of items presented per query, impacts answer quality and cognitive load. This work investigates the trade-off between query size and answer accuracy, revealing diminishing returns beyond 4-5 items per query. Crucially, we identify contextual bias in crowdworker responses – the likelihood of grouping items depends not only on their similarity but also on the other items present in the query. This structured noise contradicts assumptions made in existing noise models. Our findings underscore the need for more nuanced noise models that account for the complex interplay between items and query context in crowdsourced clustering tasks.

Cite

Text

Chen and Vinayak. "Query Design for Crowdsourced Clustering: Effect of Cognitive Overload and Contextual Bias." ICML 2024 Workshops: MFHAIA, 2024.

Markdown

[Chen and Vinayak. "Query Design for Crowdsourced Clustering: Effect of Cognitive Overload and Contextual Bias." ICML 2024 Workshops: MFHAIA, 2024.](https://mlanthology.org/icmlw/2024/chen2024icmlw-query/)

BibTeX

@inproceedings{chen2024icmlw-query,
  title     = {{Query Design for Crowdsourced Clustering: Effect of Cognitive Overload and Contextual Bias}},
  author    = {Chen, Yi and Vinayak, Ramya Korlakai},
  booktitle = {ICML 2024 Workshops: MFHAIA},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/chen2024icmlw-query/}
}