Zero-shot Sequence Labeling: Transferring Knowledge from Sentences to Tokens

Marek Rei, Anders Søgaard

    16 Citations (Scopus)

    Abstract

    Can attention-or gradient-based visualization techniques be used to infer token-level labels for binary sequence tagging problems, using networks trained only on sentence-level labels? We construct a neural network architecture based on soft attention, train it as a binary sentence classifier and evaluate against tokenlevel annotation on four different datasets. Inferring token labels from a network provides a method for quantitatively evaluating what the model is learning, along with generating useful feedback in assistance systems. Our results indicate that attention-based methods are able to predict token-level labels more accurately, compared to gradient-based methods, sometimes even rivaling the supervised oracle network.

    Original languageEnglish
    Title of host publicationProceedings, 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies : (Long Ppaers)
    Volume1
    PublisherAssociation for Computational Linguistics
    Publication date2018
    Pages293–302
    Publication statusPublished - 2018
    Event16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - New Orleans, United States
    Duration: 1 Jun 20186 Jun 2018

    Conference

    Conference16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
    Country/TerritoryUnited States
    CityNew Orleans
    Period01/06/201806/06/2018

    Fingerprint

    Dive into the research topics of 'Zero-shot Sequence Labeling: Transferring Knowledge from Sentences to Tokens'. Together they form a unique fingerprint.

    Cite this