Open Access iconOpen Access

ARTICLE

An Infrared-Visible Image Fusion Network with Channel-Switching for Low-Light Object Detection

Tianzhe Jiao, Yuming Chen, Xiaoyue Feng, Chaopeng Guo, Jie Song*

Software College, Northeastern University, Shenyang, 110819, China

* Corresponding Author: Jie Song. Email: email

Computers, Materials & Continua 2025, 85(2), 2681-2700. https://doi.org/10.32604/cmc.2025.069235

Abstract

Visible-infrared object detection leverages the day-night stable object perception capability of infrared images to enhance detection robustness in low-light environments by fusing the complementary information of visible and infrared images. However, the inherent differences in the imaging mechanisms of visible and infrared modalities make effective cross-modal fusion challenging. Furthermore, constrained by the physical characteristics of sensors and thermal diffusion effects, infrared images generally suffer from blurred object contours and missing details, making it difficult to extract object features effectively. To address these issues, we propose an infrared-visible image fusion network that realizes multimodal information fusion of infrared and visible images through a carefully designed multi-scale fusion strategy. First, we design an adaptive gray-radiance enhancement (AGRE) module to strengthen the detail representation in infrared images, improving their usability in complex lighting scenarios. Next, we introduce a channel-spatial feature interaction (CSFI) module, which achieves efficient complementarity between the RGB and infrared (IR) modalities via dynamic channel switching and a spatial attention mechanism. Finally, we propose a multi-scale enhanced cross-attention fusion (MSECA) module, which optimizes the fusion of multi-level features through dynamic convolution and gating mechanisms and captures long-range complementary relationships of cross-modal features on a global scale, thereby enhancing the expressiveness of the fused features. Experiments on the KAIST, M3FD, and FLIR datasets demonstrate that our method delivers outstanding performance in daytime and nighttime scenarios. On the KAIST dataset, the miss rate drops to 5.99%, and further to 4.26% in night scenes. On the FLIR and M3FD datasets, it achieves scores of 79.4% and 88.9%, respectively.

Keywords

Infrared–visible image fusion; channel switching; low-light object detection; cross-attention fusion

Cite This Article

APA Style
Jiao, T., Chen, Y., Feng, X., Guo, C., Song, J. (2025). An Infrared-Visible Image Fusion Network with Channel-Switching for Low-Light Object Detection. Computers, Materials & Continua, 85(2), 2681–2700. https://doi.org/10.32604/cmc.2025.069235
Vancouver Style
Jiao T, Chen Y, Feng X, Guo C, Song J. An Infrared-Visible Image Fusion Network with Channel-Switching for Low-Light Object Detection. Comput Mater Contin. 2025;85(2):2681–2700. https://doi.org/10.32604/cmc.2025.069235
IEEE Style
T. Jiao, Y. Chen, X. Feng, C. Guo, and J. Song, “An Infrared-Visible Image Fusion Network with Channel-Switching for Low-Light Object Detection,” Comput. Mater. Contin., vol. 85, no. 2, pp. 2681–2700, 2025. https://doi.org/10.32604/cmc.2025.069235



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 1943

    View

  • 1159

    Download

  • 0

    Like

Share Link