Interrater disagreement resolution: A systematic procedure to reach consensus in annotation tasks

Autor: Oortwijn, Y., Ossenkoppele, T., Betti, A., Belz, A., Agarwal, S., Graham, Y., Reiter, E., Shimorina, A.
Přispěvatelé: ILLC (FGw), Logic and Language (ILLC, FNWI/FGw), ILLC (FNWI/FGw)
Jazyk: angličtina
Rok vydání: 2021
Zdroj: Human Evaluation of NLP Systems (HumEval): EACL 2021 : proceedings of the workshop : April 19, 2021
Human Evaluation of NLP Systems (HumEval)
Popis: We present a systematic procedure for interrater disagreement resolution. The procedure is general, but of particular use in multiple-annotator tasks geared towards ground truth construction. We motivate our proposal by arguing that, barring cases in which the researchers’ goal is to elicit different viewpoints, interrater disagreement is a sign of poor quality in the design or the description of a task. Consensus among annotators, we maintain, should be striven for, through a systematic procedure for disagreement resolution such as the one we describe.
Databáze: OpenAIRE