Zur Kurzanzeige

dc.contributor.authorKlie, Jan-Christoph
dc.contributor.authorEckart de Castilho, Richard
dc.contributor.authorGurevych, Iryna
dc.date.accessioned2023-09-07T21:45:11Z
dc.date.available2023-09-07T21:45:11Z
dc.date.issued2023-09-07
dc.identifier.urihttps://tudatalib.ulb.tu-darmstadt.de/handle/tudatalib/3939
dc.identifier.urihttps://doi.org/10.48328/tudatalib-1220
dc.descriptionThis is the accompanying data for the paper "Analyzing Dataset Annotation Quality Management in the Wild". Data quality is crucial for training accurate, unbiased, and trustworthy machine learning models and their correct evaluation. Recent works, however, have shown that even popular datasets used to train and evaluate state-of-the-art models contain a non-negligible amount of erroneous annotations, bias or annotation artifacts. There exist best practices and guidelines regarding annotation projects. But to the best of our knowledge, no large-scale analysis has been performed as of yet on how quality management is actually conducted when creating natural language datasets and whether these recommendations are followed. Therefore, we first survey and summarize recommended quality management practices for dataset creation as described in the literature and provide suggestions on how to apply them. Then, we compile a corpus of 591 scientific publications introducing text datasets and annotate it for quality-related aspects, such as annotator management, agreement, adjudication or data validation. Using these annotations, we then analyze how quality management is conducted in practice. We find that a majority of the annotated publications apply good or very good quality management. However, we deem the effort of 30% of the works as only subpar. Our analysis also shows common errors, especially with using inter-annotator agreement and computing annotation error rates.de_DE
dc.relationIsSupplementTo;arXiv;2307.08153
dc.relationIsReferencedBy;URL;https://github.com/UKPLab/arxiv2023-qanno
dc.rightsCreative Commons Attribution-NonCommercial 4.0
dc.rights.urihttps://creativecommons.org/licenses/by-nc/4.0/
dc.subjectannotationde_DE
dc.subjectquality managementde_DE
dc.subjectnlpde_DE
dc.subject.classification4.43-04 Künstliche Intelligenz und Maschinelle Lernverfahrende_DE
dc.subject.classification4.43-05 Bild- und Sprachverarbeitung, Computergraphik und Visualisierung, Human Computer Interaction, Ubiquitous und Wearable Computing
dc.subject.ddc004
dc.titleAnalyzing Dataset Annotation Quality Management in the Wildde_DE
dc.typeDatasetde_DE
tud.unitTUDa
tud.history.classificationVersion=2016-2020;409-05 Interaktive und intelligente Systeme, Bild- und Sprachverarbeitung, Computergraphik und Visualisierung


Dateien zu dieser Ressource

Thumbnail
Thumbnail

Der Datensatz erscheint in:

Zur Kurzanzeige

Creative Commons Attribution-NonCommercial 4.0
Solange nicht anders angezeigt, wird die Lizenz wie folgt beschrieben: Creative Commons Attribution-NonCommercial 4.0