MEAformer: Multi-modal Entity Alignment Transformer for Meta Modality Hybrid

15Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Multi-modal entity alignment (MMEA) aims to discover identical entities across different knowledge graphs (KGs) whose entities are associated with relevant images. However, current MMEA algorithms rely on KG-level modality fusion strategies for multi-modal entity representation, which ignores the variations of modality preferences of different entities, thus compromising robustness against noise in modalities such as blurry images and relations. This paper introduces MEAformer, a mlti-modal entity alignment transformer approach for meta modality hybrid, which dynamically predicts the mutual correlation coefficients among modalities for more fine-grained entity-level modality fusion and alignment. Experimental results demonstrate that our model not only achieves SOTA performance in multiple training scenarios, including supervised, unsupervised, iterative, and low-resource settings, but also has a limited number of parameters, efficient runtime, and interpretability. Our code is available at https://github.com/zjukg/MEAformer.

Cite

CITATION STYLE

APA

Chen, Z., Chen, J., Zhang, W., Guo, L., Fang, Y., Huang, Y., … Chen, H. (2023). MEAformer: Multi-modal Entity Alignment Transformer for Meta Modality Hybrid. In MM 2023 - Proceedings of the 31st ACM International Conference on Multimedia (pp. 3317–3327). Association for Computing Machinery, Inc. https://doi.org/10.1145/3581783.3611786

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free