Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Long-term Annotators: A Supervised Label Aggregation Baseline (2311.14709v1)

Published 15 Nov 2023 in cs.HC and cs.LG

Abstract: Relying on crowdsourced workers, data crowdsourcing platforms are able to efficiently provide vast amounts of labeled data. Due to the variability in the annotation quality of crowd workers, modern techniques resort to redundant annotations and subsequent label aggregation to infer true labels. However, these methods require model updating during the inference, posing challenges in real-world implementation. Meanwhile, in recent years, many data labeling tasks have begun to require skilled and experienced annotators, leading to an increasing demand for long-term annotators. These annotators could leave substantial historical annotation records on the crowdsourcing platforms, which can benefit label aggregation, but are ignored by previous works. Hereby, in this paper, we propose a novel label aggregation technique, which does not need any model updating during inference and can extensively explore the historical annotation records. We call it SuperLA, a Supervised Label Aggregation method. Inside this model, we design three types of input features and a straightforward neural network structure to merge all the information together and subsequently produce aggregated labels. Based on comparison experiments conducted on 22 public datasets and 11 baseline methods, we find that SuperLA not only outperforms all those baselines in inference performance but also offers significant advantages in terms of efficiency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Haoyu Liu (49 papers)
  2. Fei Wang (574 papers)
  3. Minmin Lin (6 papers)
  4. Runze Wu (28 papers)
  5. Renyu Zhu (17 papers)
  6. Shiwei Zhao (6 papers)
  7. Kai Wang (624 papers)
  8. Tangjie Lv (35 papers)
  9. Changjie Fan (79 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.