ORIGINAL RESEARCH article

Front. Phys.

Sec. Optics and Photonics

Volume 13 - 2025 | doi: 10.3389/fphy.2025.1514476

Infrared and visible image fusion based on multi-scale transform and sparse low-rank representation

Provisionally accepted
Yangkun  ZouYangkun Zou1,2Jiande  WuJiande Wu1,3Bo  YeBo Ye4,5*Honggui  CaoHonggui Cao4,5Jiqi  FengJiqi Feng6Zijie  WanZijie Wan4,5Shaoda  YinShaoda Yin4,5
  • 1School of Information Science and Engineering, Yunnan University, Kunming, Yunnan Province, China
  • 2Faculty of Civil Aviation and Aeronautics, Kunming University of Science and Technology, Kunming, Yunnan Province, China
  • 3Yunnan Key Laboratory of Intelligent Systems and Computing, Yunnan University, Kunming, China
  • 4Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, Yunnan Province, China
  • 5Yunnan Key Laboratory of Intelligent Control and Application, Kunming University of Science and Technology, Kunming, China
  • 6Guangxi Huasheng new material Co., LTD, Fangchenggang, China

The final, formatted version of the article will be published soon.

In order to improve the fusion of infrared and visible images, a novel and effective fusion method is proposed based on multi-scale transform and sparse low-rank representation in this paper. Visible and infrared images are first decomposed to obtain their low-pass and high-pass bands by Laplacian pyramid (LP). Second, low-pass bands are represented with some sparse and low-rank coefficients. In order to improve the computational efficiency and learn a universal dictionary, low-pass bands are separated into several image patches using a sliding window prior to sparse and low rank representation. The low-pass and high-pass bands are then fused by particular fusion rules. The max-absolute rule is used to fuse the high-pass bands, and max-L1 norm rule is utilized to fuse the low-pass bands. Finally, an inverse LP is performed to acquire the fused image. We conduct experiments on three datasets and use 13 metrics to thoroughly and impartially validate our method. The results demonstrate that the proposed fusion framework can effectively preserve the characteristics of source images, and exhibits superior stability across various image pairs and metrics.

Keywords: image fusion, Multi-scale transform, sparse representation, Low-rank representation, Infrared image, Visible image

Received: 21 Oct 2024; Accepted: 15 May 2025.

Copyright: © 2025 Zou, Wu, Ye, Cao, Feng, Wan and Yin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Bo Ye, Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, 650504, Yunnan Province, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.