MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences

62Citations
Citations of this article
146Readers
Mendeley users who have this article in their library.

Abstract

Human communication is multimodal in nature; it is through multiple modalities such as language, voice, and facial expressions, that opinions and emotions are expressed. Data in this domain exhibits complex multi-relational and temporal interactions. Learning from this data is a fundamentally challenging research problem. In this paper, we propose Modal-Temporal Attention Graph (MTAG). MTAG is an interpretable graph-based neural model that provides a suitable framework for analyzing multimodal sequential data. We first introduce a procedure to convert unaligned multimodal sequence data into a graph with heterogeneous nodes and edges that captures the rich interactions across modalities and through time. Then, a novel graph fusion operation, called MTAG fusion, along with a dynamic pruning and read-out technique, is designed to efficiently process this modal-temporal graph and capture various interactions. By learning to focus only on the important interactions within the graph, MTAG achieves state-of-the-art performance on multimodal sentiment analysis and emotion recognition benchmarks, while utilizing significantly fewer model parameters.

Cite

CITATION STYLE

APA

Yang, J., Wang, Y., Yi, R., Zhu, Y., Rehman, A., Zadeh, A., … Morency, L. P. (2021). MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1009–1021). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.79

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free