Journal of Computer Applications ›› 2025, Vol. 45 ›› Issue (2): 594-600.DOI: 10.11772/j.issn.1001-9081.2024030302

• Multimedia computing and computer simulation • Previous Articles    

Multi-focus image fusion network with cascade fusion and enhanced reconstruction

Benchen YANG(), Haoran LI, Haibo JIN   

  1. School of Software,Liaoning Technical University,Huludao Liaoning 125105,China
  • Received:2024-03-21 Revised:2024-06-10 Accepted:2024-06-13 Online:2024-07-31 Published:2025-02-10
  • Contact: Benchen YANG
  • About author:LI Haoran, born in 2000, M. S. candidate. His research interests include multi-focus image fusion, image super-resolution reconstruction, image enhancement.
    JIN Haibo, born in 1983, Ph. D., associate professor. His research interests include deep learning, computer vision, reliability analysis of complex systems.
  • Supported by:
    National Natural Science Foundation of China(62173171)

级联融合与增强重建的多聚焦图像融合网络

杨本臣(), 李浩然, 金海波   

  1. 辽宁工程技术大学 软件学院,辽宁 葫芦岛 125105
  • 通讯作者: 杨本臣
  • 作者简介:李浩然(2000—),男,辽宁铁岭人,硕士研究生,CCF会员,主要研究方向:多聚焦图像融合、图像超分辨率重建、图像增强
    金海波(1983—),男,辽宁沈阳人,副教授,博士,CCF会员,主要研究方向:深度学习、计算机视觉、复杂系统可靠性分析。
  • 基金资助:
    国家自然科学基金资助项目(62173171)

Abstract:

Aiming at the problem of semi-focus images caused by improper focusing of far and near visual fields during digital image shooting, a multi-focus image fusion Network with Cascade fusion and enhanced reconstruction (CasNet) was proposed. Firstly, a cascade sampling module was constructed to calculate and merge the residuals of feature maps sampled at different depths for efficient utilization of focused features at different scales. Secondly, a lightweight multi-head self-attention mechanism was improved to perform dimensional residual calculation on feature maps for feature enhancement of the image and make the feature maps present better distribution in different dimensions. Thirdly, convolution channel attention stacking was used to complete feature reconstruction. Finally, interval convolution was used for up- and down-sampling during the sampling process, so as to retain more original image features. Experimental results demonstrate that CasNet achieves better results in metrics such as Average Gradient (AG) and Gray-Level Difference (GLD) on multi-focus image benchmark test sets Lytro, MFFW, grayscale, and MFI-WHU compared to popular methods such as SESF-Fuse (Spatially Enhanced Spatial Frequency-based Fusion) and U2Fusion (Unified Unsupervised Fusion network).

Key words: multi-focus image fusion, Deep Neural Network (DNN), feature reconstruction, feature enhancement, attention

摘要:

针对数字图像拍摄过程中因远近视野聚焦不当所导致的半聚焦图像问题,提出一种级联融合与增强重建的多聚焦图像融合网络(CasNet)。首先,构建级联采样模块对不同深度采样特征图的残差进行计算与合并,从而高效利用不同尺度下的聚焦特征;其次,改进轻量化多头自注意力机制以计算特征图的维度残差,从而完成图像的特征增强,并使特征图在不同维度上呈现更优分布;再次,使用卷积通道注意力堆叠完成特征重建;最后,在采样过程中使用分隔卷积进行上下采样,从而保留更多的图像原有特征。实验结果表明,在多聚焦图像基准测试集Lytro、MFFW、grayscale和MFI-WHU上,CasNet相较于SESF-Fuse(Spatially Enhanced Spatial Frequency-based Fusion)和U2Fusion(Unified Unsupervised Fusion network)等热门方法在平均梯度(AG)、灰度级差(GLD)等指标上都取得了较好的结果。

关键词: 多聚焦图像融合, 深度神经网络, 特征重建, 特征增强, 注意力

CLC Number: