Creative Commons Attribution 3.0 Unported license
Event detection is still a difficult task due to the complexity and the ambiguity of such entities. On the one hand, we observe a low inter-annotator agreement among experts when annotating events, disregarding the multitude of existing annotation guidelines and their numerous revisions. On the other hand, event extraction systems have a lower measured performance in terms of F1-score compared to other types of entities such as people or locations. In this paper we study the consistency and completeness of expert-annotated datasets for events and time expressions. We propose a data-agnostic validation methodology of such datasets in terms of consistency and completeness. Furthermore, we combine the power of crowds and machines to correct and extend expert-annotated datasets of events. We show the benefit of using crowd-annotated events to train and evaluate a state-of-the-art event extraction system. Our results show that the crowd-annotated events increase the performance of the system by at least 5.3%.
@InProceedings{inel_et_al:OASIcs.LDK.2019.12,
author = {Inel, Oana and Aroyo, Lora},
title = {{Validation Methodology for Expert-Annotated Datasets: Event Annotation Case Study}},
booktitle = {2nd Conference on Language, Data and Knowledge (LDK 2019)},
pages = {12:1--12:15},
series = {Open Access Series in Informatics (OASIcs)},
ISBN = {978-3-95977-105-4},
ISSN = {2190-6807},
year = {2019},
volume = {70},
editor = {Eskevich, Maria and de Melo, Gerard and F\"{a}th, Christian and McCrae, John P. and Buitelaar, Paul and Chiarcos, Christian and Klimek, Bettina and Dojchinovski, Milan},
publisher = {Schloss Dagstuhl -- Leibniz-Zentrum f{\"u}r Informatik},
address = {Dagstuhl, Germany},
URL = {https://drops.dagstuhl.de/entities/document/10.4230/OASIcs.LDK.2019.12},
URN = {urn:nbn:de:0030-drops-103762},
doi = {10.4230/OASIcs.LDK.2019.12},
annote = {Keywords: Crowdsourcing, Human-in-the-Loop, Event Extraction, Time Extraction}
}