SGMEA: Structure-Guided Multimodal Entity Alignment

Jingwei Cheng, Mingxiao Guo, Fu Zhang


Abstract
Multimodal Entity Alignment (MMEA) aims to identify equivalent entities across different multimodal knowledge graphs (MMKGs) by integrating structural information, entity attributes, and visual data, thereby promoting knowledge sharing and deep multimodal data integration. However, existing methods often overlook the deeper connections between multimodal data. They primarily focus on the interactions between neighboring entities in the structural modality while neglecting the interactions between entities in the visual and attribute modalities. To address this, we propose a structure-guided multimodal entity alignment method (SGMEA), which prioritizes structural information from knowledge graphs to enhance the visual and attribute modalities. By fusing multimodal representations, SGMEA improves the accuracy of entity alignment. Experimental results demonstrate that SGMEA achieves stateof-the-art performance across multiple datasets, validating its effectiveness and superiority in practical applications.
Anthology ID:
2025.coling-main.525
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7851–7861
Language:
URL:
https://aclanthology.org/2025.coling-main.525/
DOI:
Bibkey:
Cite (ACL):
Jingwei Cheng, Mingxiao Guo, and Fu Zhang. 2025. SGMEA: Structure-Guided Multimodal Entity Alignment. In Proceedings of the 31st International Conference on Computational Linguistics, pages 7851–7861, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
SGMEA: Structure-Guided Multimodal Entity Alignment (Cheng et al., COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.525.pdf