Learning to parse with IAA-weighted loss

Hector Martinez Alonso, Barbara Plank, Anders Søgaard

1 Citation (Scopus)

Abstract

Natural language processing (NLP) annotation projects employ guidelines to maximize inter-annotator agreement (IAA), and models are estimated assuming that there is one single ground truth. However, not all disagreement is noise, and in fact some of it may contain valuable linguistic information. We integrate such information in the training of a cost-sensitive dependency parser. We introduce five different factorizations of IAA and the corresponding loss functions, and evaluate these across six different languages. We obtain robust improvements across the board using a factorization that considers dependency labels and directionality. The best method-dataset combination reaches an average overall error reduction of 6.4% in labeled attachment score.

Original languageEnglish
Title of host publicationProceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies : NAACL 2015
Number of pages5
PublisherAssociation for Computational Linguistics
Publication date2015
Pages1357-1361
ISBN (Print)978-1-941643-49-5
Publication statusPublished - 2015

Fingerprint

Dive into the research topics of 'Learning to parse with IAA-weighted loss'. Together they form a unique fingerprint.

Cite this