Structured Fusion Attention Network for Image Super-Resolution Reconstruction

To improve the extraction ability of image features, reduce the complexity of model parameters, and enhance the reconstruction effect of image super-resolution (SR), a structured fusion attention network (SFAN) is proposed. Firstly, the deep convolution method is used to extract shallow features fro...

Full description

Bibliographic Details
Main Authors: Yaonan Dai, Jiuyang Yu, Tianhao Hu, Yang Lu, Xiaotao Zheng
Format: Article
Language:English
Published: IEEE 2022-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9737465/
_version_ 1818356302908227584
author Yaonan Dai
Jiuyang Yu
Tianhao Hu
Yang Lu
Xiaotao Zheng
author_facet Yaonan Dai
Jiuyang Yu
Tianhao Hu
Yang Lu
Xiaotao Zheng
author_sort Yaonan Dai
collection DOAJ
description To improve the extraction ability of image features, reduce the complexity of model parameters, and enhance the reconstruction effect of image super-resolution (SR), a structured fusion attention network (SFAN) is proposed. Firstly, the deep convolution method is used to extract shallow features from low-resolution images, and different residual attention modules are considered to improve the structured residual of the encoder to extract more image features. Secondly, the features output by the encoder are refined, and the spatial attention module and the channel attention module are reorganized according to an improved fusion attention method to provide better input features for PixelShuffle, thus achieving the effect of reconstructing the decoder. Finally, through adding low-frequency inputs and network predictions, the input image is directly interpolated into the target, thereby accelerating the convergence of the network&#x2019;s high-frequency residual and improving the image reconstruction effect. Under the condition of reconstruction magnification of <inline-formula> <tex-math notation="LaTeX">$\times 2$ </tex-math></inline-formula>, <inline-formula> <tex-math notation="LaTeX">$\times 3$ </tex-math></inline-formula>, <inline-formula> <tex-math notation="LaTeX">$\times 4$ </tex-math></inline-formula>, SFAN is compared with some of the most advanced SR networks in the public data sets of Set5, Set14, BSD100, Urban100 and Manga 109. The experimental results show that SFAN has the best PSNR and SSIM values with low model parameters, thus proving that SFAN can achieve a good balance between the performance of SR and the complexity of parameters.
first_indexed 2024-12-13T19:55:04Z
format Article
id doaj.art-9883a2db2347410781421047aac19af3
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-13T19:55:04Z
publishDate 2022-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-9883a2db2347410781421047aac19af32022-12-21T23:33:19ZengIEEEIEEE Access2169-35362022-01-0110318963190610.1109/ACCESS.2022.31605589737465Structured Fusion Attention Network for Image Super-Resolution ReconstructionYaonan Dai0Jiuyang Yu1Tianhao Hu2Yang Lu3Xiaotao Zheng4Hubei Provincial Engineering Technology Research Center of Green Chemical Equipment, School of Mechanical and Electrical Engineering, Wuhan Institute of Technology, Wuhan, ChinaHubei Provincial Engineering Technology Research Center of Green Chemical Equipment, School of Mechanical and Electrical Engineering, Wuhan Institute of Technology, Wuhan, ChinaHubei Provincial Engineering Technology Research Center of Green Chemical Equipment, School of Mechanical and Electrical Engineering, Wuhan Institute of Technology, Wuhan, ChinaSchool of Machinery and Automation, Wuhan University of Science and Technology, Wuhan, ChinaHubei Provincial Engineering Technology Research Center of Green Chemical Equipment, School of Mechanical and Electrical Engineering, Wuhan Institute of Technology, Wuhan, ChinaTo improve the extraction ability of image features, reduce the complexity of model parameters, and enhance the reconstruction effect of image super-resolution (SR), a structured fusion attention network (SFAN) is proposed. Firstly, the deep convolution method is used to extract shallow features from low-resolution images, and different residual attention modules are considered to improve the structured residual of the encoder to extract more image features. Secondly, the features output by the encoder are refined, and the spatial attention module and the channel attention module are reorganized according to an improved fusion attention method to provide better input features for PixelShuffle, thus achieving the effect of reconstructing the decoder. Finally, through adding low-frequency inputs and network predictions, the input image is directly interpolated into the target, thereby accelerating the convergence of the network&#x2019;s high-frequency residual and improving the image reconstruction effect. Under the condition of reconstruction magnification of <inline-formula> <tex-math notation="LaTeX">$\times 2$ </tex-math></inline-formula>, <inline-formula> <tex-math notation="LaTeX">$\times 3$ </tex-math></inline-formula>, <inline-formula> <tex-math notation="LaTeX">$\times 4$ </tex-math></inline-formula>, SFAN is compared with some of the most advanced SR networks in the public data sets of Set5, Set14, BSD100, Urban100 and Manga 109. The experimental results show that SFAN has the best PSNR and SSIM values with low model parameters, thus proving that SFAN can achieve a good balance between the performance of SR and the complexity of parameters.https://ieeexplore.ieee.org/document/9737465/Image super-resolutionstructured residualfusion attentionlow model parametersSFAN
spellingShingle Yaonan Dai
Jiuyang Yu
Tianhao Hu
Yang Lu
Xiaotao Zheng
Structured Fusion Attention Network for Image Super-Resolution Reconstruction
IEEE Access
Image super-resolution
structured residual
fusion attention
low model parameters
SFAN
title Structured Fusion Attention Network for Image Super-Resolution Reconstruction
title_full Structured Fusion Attention Network for Image Super-Resolution Reconstruction
title_fullStr Structured Fusion Attention Network for Image Super-Resolution Reconstruction
title_full_unstemmed Structured Fusion Attention Network for Image Super-Resolution Reconstruction
title_short Structured Fusion Attention Network for Image Super-Resolution Reconstruction
title_sort structured fusion attention network for image super resolution reconstruction
topic Image super-resolution
structured residual
fusion attention
low model parameters
SFAN
url https://ieeexplore.ieee.org/document/9737465/
work_keys_str_mv AT yaonandai structuredfusionattentionnetworkforimagesuperresolutionreconstruction
AT jiuyangyu structuredfusionattentionnetworkforimagesuperresolutionreconstruction
AT tianhaohu structuredfusionattentionnetworkforimagesuperresolutionreconstruction
AT yanglu structuredfusionattentionnetworkforimagesuperresolutionreconstruction
AT xiaotaozheng structuredfusionattentionnetworkforimagesuperresolutionreconstruction