Deep multi-task learning with low level tasks supervised at lower layers

Anders Søgaard, Yoav Goldberg

173 Citationer (Scopus)

Abstract

In all previous work on deep multi-task learning we are aware of, all task supervisions are on the same (outermost) layer. We present a multi-task learning architecture with deep bi-directional RNNs, where different tasks supervision can happen at different layers. We present experiments in syntactic chunking and CCG supertagging, coupled with the additional task of POS-tagging. We show that it is consistently better to have POS supervision at the innermost rather than the outermost layer. We argue that this is because "lowlevel" tasks are better kept at the lower layers, enabling the higher-level tasks to make use of the shared representation of the lower-level tasks. Finally, we also show how this architecture can be used for domain adaptation.

OriginalsprogEngelsk
TitelProceedings of the 54th Annual Meeting of the Association for Computational Linguistics
Antal sider5
Vol/bind2
ForlagAssociation for Computational Linguistics
Publikationsdato2016
Sider231-235
ISBN (Elektronisk)978-1-945626-01-2
StatusUdgivet - 2016
Begivenhed54th Annual Meeting of the Association for Computational Linguistics - Berlin, Tyskland
Varighed: 7 aug. 201612 aug. 2016
Konferencens nummer: 54

Konference

Konference54th Annual Meeting of the Association for Computational Linguistics
Nummer54
Land/OmrådeTyskland
ByBerlin
Periode07/08/201612/08/2016

Fingeraftryk

Dyk ned i forskningsemnerne om 'Deep multi-task learning with low level tasks supervised at lower layers'. Sammen danner de et unikt fingeraftryk.

Citationsformater