Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm

Bjarke Felbo, Alan Mislove, Anders Søgaard, Iyad Rahwan, Sune Lehmann

144 Citations (Scopus)

Abstract

NLP tasks are often limited by scarcity of
manually annotated data. In social media
sentiment analysis and related tasks,
researchers have therefore used binarized
emoticons and specific hashtags as forms
of distant supervision. Our paper shows
that by extending the distant supervision
to a more diverse set of noisy labels, the
models can learn richer representations.
Through emoji prediction on a dataset of
1246 million tweets containing one of 64
common emojis we obtain state-of-theart
performance on 8 benchmark datasets
within emotion, sentiment and sarcasm detection
using a single pretrained model.
Our analyses confirm that the diversity of
our emotional labels yield a performance
improvement over previous distant supervision
approaches.
Original languageEnglish
Title of host publicationProceedings of the 2017 Conference on Empirical Methods in Natural Language Processing
Number of pages11
PublisherAssociation for Computational Linguistics
Publication date2017
Pages615–1625
Publication statusPublished - 2017
Event2017 Conference on Empirical Methods in Natural Language Processing - Copemhagen, Denmark
Duration: 9 Sept 201711 Sept 2017

Conference

Conference2017 Conference on Empirical Methods in Natural Language Processing
Country/TerritoryDenmark
CityCopemhagen
Period09/09/201711/09/2017

Fingerprint

Dive into the research topics of 'Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm'. Together they form a unique fingerprint.

Cite this