Back to articles
Special issue: ACDM 2025
Volume: 70 | Article ID: 010415
Image
GMFDE: Gated Multi-Modal Fusion with Distillation Enhancement for Entity Alignment
  DOI :  10.2352/J.ImagingSci.Technol.2026.70.1.010415  Published OnlineJanuary 2026
Abstract
Abstract

In recent years, multi-modal knowledge graphs (MMKGs) have emerged to enhance the representation of real-world entities through structural, textual, and visual features. However, the inherent heterogeneity among different modalities poses significant challenges for entity alignment across KGs. In this study, we introduce GMFDE, an innovative framework for multi-modal entity alignment that integrates a gated residual fusion mechanism with a knowledge distillation strategy. The fusion module adaptively balances and refines modality-specific features while the distillation component enables unimodal encoders to learn complementary information from the fused multi-modal representation, promoting consistency across modalities. Extensive experiments on both bilingual and cross-KG datasets demonstrate that GMFDE achieves superior performance compared with existing leading methods, particularly excelling in settings with limited alignment seeds.

Subject Areas :
Views 22
Downloads 1
 articleview.views 22
 articleview.downloads 1
  Cite this article 

Miaomiao Li, Rongyan Yuan, Zhijiang Li, "GMFDE: Gated Multi-Modal Fusion with Distillation Enhancement for Entity Alignmentin Journal of Imaging Science and Technology,  2026,  pp 1 - 11,  https://doi.org/10.2352/J.ImagingSci.Technol.2026.70.1.010415

 Copy citation
  Copyright statement 
Copyright © Society for Imaging Science and Technology 2026
  Article timeline 
  • received July 2025
  • accepted December 2025
  • PublishedJanuary 2026

Preprint submitted to:
  Login or subscribe to view the content