Adversarial Evaluation of Multimodal Machine Translation

    Abstract

    The promise of combining vision and language in multimodal machine translation is that systems will produce better translations by leveraging the image data. However, inconsistent results have lead to uncertainty about whether the images actually improve translation quality. We present an adversarial evaluation method to directly examine the utility of the image data in this task. Our evaluation measures whether multimodal translation systems perform better given either the congruent image or a random incongruent image, in addition to the correct source language sentence. We find that two out of three publicly available systems are sensitive to this perturbation of the data, and recommend that all systems pass this evaluation in the future.

    Original languageEnglish
    Title of host publicationProceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
    Number of pages5
    Publication date2018
    Pages2974-2978
    Publication statusPublished - 2018

    Fingerprint

    Dive into the research topics of 'Adversarial Evaluation of Multimodal Machine Translation'. Together they form a unique fingerprint.

    Cite this