Abstract
Sentiment analysis models often use ratings as labels, assuming that these ratings reflect the sentiment of the accompanying text. We investigate (i) whether human readers can infer ratings from review text, (ii) how human performance compares to a regression model, and (iii) whether model performance is affected by the rating "source" (i.e. original author vs. annotator). We collect IMDb movie reviews with author-provided ratings, and have them re-annotated by crowdsourced and trained annotators. Annotators reproduce the original ratings better than a model, but are still far off in more than 5% of the cases. Models trained on annotator-labels outperform those trained on author-labels, questioning the usefulness of author-rated reviews as training data for sentiment analysis.
Original language | English |
---|---|
Title of host publication | 2015 Conference on Empirical Methods for Natural Language Processing |
Number of pages | 6 |
Place of Publication | Lisbon, Portugal |
Publisher | Association for Computational Linguistics |
Publication date | 2015 |
Pages | 2527-2532 |
ISBN (Print) | 978-1-941643-32-7 |
Publication status | Published - 2015 |