Open Peer Review. Open Publishing. Open Access. Open Discussion. Open Directory. Open Recommendations. Open API. Open Source.
Multimodal Sentiment Analysis To Explore the Structure of Emotions
Anthony Hu, Seth Flaxman
Feb 15, 2018 (modified: Feb 15, 2018)ICLR 2018 Conference Blind Submissionreaders: everyoneShow Bibtex
Abstract:We propose a novel approach to multimodal sentiment analysis using deep neural
networks combining visual recognition and natural language processing. Our
goal is different than the standard sentiment analysis goal of predicting whether
a sentence expresses positive or negative sentiment; instead, we aim to infer the
latent emotional state of the user. Thus, we focus on predicting the emotion word
tags attached by users to their Tumblr posts, treating these as “self-reported emotions.”
We demonstrate that our multimodal model combining both text and image
features outperforms separate models based solely on either images or text. Our
model’s results are interpretable, automatically yielding sensible word lists associated
with emotions. We explore the structure of emotions implied by our model
and compare it to what has been posited in the psychology literature, and validate
our model on a set of images that have been used in psychology studies. Finally,
our work also provides a useful tool for the growing academic study of images—
both photographs and memes—on social networks.
Enter your feedback below and we'll get back to you as soon as possible.