Abstract
Multimodal sentiment analysis has recently received an increasing interest. However, most methods have considered that text and image modalities are always available at test time. This assumption is often violated in real environments (e.g. social media) since users do not always publish a text with an image. In this paper we propose a method based on a multitask framework to combine multimodal information when it is available, while being able to handle the cases where a modality is missing. Our model contains one classifier for analyzing the text, another for analyzing the image, and another performing the prediction by fusing both modalities. In addition to offer a solution to the problem of a missing modality, our experiments show that this multitask framework improves generalization by acting as a regularization mechanism. We also demonstrate that the model can handle a missing modality at training time, thus being able to be trained with image-only and text-only examples.
Author supplied keywords
Cite
CITATION STYLE
Fortin, M. P., & Chaib-Draa, B. (2019). Multimodal Sentiment Analysis: A Multitask Learning Approach. In International Conference on Pattern Recognition Applications and Methods (Vol. 1, pp. 368–376). Science and Technology Publications, Lda. https://doi.org/10.5220/0007313503680376
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.