Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (2): 300-311.DOI: 10.11996/JG.j.2095-302X.2025020300

• Image Processing and Computer Vision • Previous Articles     Next Articles

MSFAFuse: sar and optical image fusion model based on multi-scale feature information and attention mechanism

PAN Shuyan(), LIU Liqun()   

  1. College of Information Science and Technology, Gansu Agricultural University, Lanzhou Gansu 730070, China
  • Received:2024-07-29 Accepted:2024-12-04 Online:2025-04-30 Published:2025-04-24
  • Contact: LIU Liqun
  • About author:First author contact:

    PAN Shuyan (2002-), master student. His main research interests cover deep learning and digital image processing.
    E-mail:pansy@st.gsau.edu.cn

  • Supported by:
    National Natural Science Foundation of China(32460440);Gansu Provincial University Teacher Innovation Fund Project(2023A-051);Young Supervisor Fund of Gansu Agricultural University(GAU-QDFC-2020-08);Gansu Science and Technology Plan(20JR5RA032)

Abstract:

In response to the issue that remote sensing images obtained from a single imaging principle cannot provide rich information, heterogeneous remote sensing image fusion technology has emerged. The imaging of synthetic aperture radar images is not affected by factors such as clouds and weather, but lacks visual observation capability. The imaging of optical images is susceptible to harsh environments, but offer direct viewing effects and target interpretation capabilities. Integrating the two can fully leverage their respective advantages to obtain high-quality images that contain more feature information and possess visual observation capabilities. To fully utilize the different scale features of heterogeneous images, a SAR and optical image fusion model based on multi-scale feature information and attention mechanism (MSFAFuse) was proposed. Firstly, robust feature downsampling was introduced to form the feature extraction part, obtaining multi-scale features corresponding to heterogeneous images. Secondly, a feature enhancement module was employed to enhance the structural features and salient regional features in heterogeneous features at different scales. Then, using a dual-branch fusion module guided by feature information and L1-Norm, was used to fuse the obtained heterogeneous multi-scale features pairwise according to scale. Finally, the fusion results of different scales were input into the image reconstruction module for image reconstruction, resulting in the final fused image. Experiments have shown that the MSFAFuse model can smoothly enhance prominent features while preserving more details and structural information. Compared with existing fusion methods, the model has shown better performance across 10 different indicators. This demonstrated that the MSFAFuse model can effectively fuse optical images and SAR images, providing new insights for the development of their fusion and contributing to the advancement of future remote sensing image fusion technologies.

Key words: image fusion, multi-scale feature, Efficient additive attention, remote sense, deep learning

CLC Number: