Back to articles
Regular Article
Volume: 69 | Article ID: 020505
Image
Illumination-guided with Crossmodal Transformer Fusion for RGB-T Object Detection
  DOI :  10.2352/J.ImagingSci.Technol.2025.69.2.020505  Published OnlineMarch 2025
Abstract
Abstract

Object detection in varying traffic scenes presents significant challenges in real-world applications. Thermal image utilization is acknowledged as a beneficial approach to enhance RGB image detection, especially in suboptimal lighting conditions. However, harnessing the combined potential of RGB and thermal images remains a formidable task. We tackle this by implementing an illumination-guided adaptive information fusion technique across both data types. Thus, we propose the illumination-guided with crossmodal attention transformer fusion (ICATF), a novel object detection framework that skillfully integrates features from RGB and thermal data. Further, an illumination-guided module is developed to adapt features to current lighting conditions, steering the learning process towards the most informative data fusion. Then, we incorporate frequency domain convolutions within the network’s backbone to assimilate spectral context and derive more nuanced features. In addition, we fuse the differential modality features for multispectral pedestrian detection with illumination-guided feature weights and transformer fusion architecture. Our method achieves state-of-the-art by experimental results on multispectral detection datasets, including FLIR-aligned, LLVIP, and KAIST.

Subject Areas :
Views 30
Downloads 16
 articleview.views 30
 articleview.downloads 16
  Cite this article 

Ruilin Xie, Sheng Jiang, Yiming Bie, Miaolei Xia, "Illumination-guided with Crossmodal Transformer Fusion for RGB-T Object Detectionin Journal of Imaging Science and Technology,  2025,  pp 1 - 11,  https://doi.org/10.2352/J.ImagingSci.Technol.2025.69.2.020505

 Copy citation
  Copyright statement 
Copyright © Society for Imaging Science and Technology 2025
  Article timeline 
  • received May 2024
  • accepted October 2024
  • PublishedMarch 2025

Preprint submitted to:
  Login or subscribe to view the content