Spatial Channel Attention for Deep Convolutional Neural Networks

Recently, the attention mechanism combining spatial and channel information has been widely used in various deep convolutional neural networks (CNNs), proving its great potential in improving model performance. However, this usually uses 2D global pooling operations to compress spatial information o...

Full description

Bibliographic Details
Main Authors: Tonglai Liu, Ronghai Luo, Longqin Xu, Dachun Feng, Liang Cao, Shuangyin Liu, Jianjun Guo
Format: Article
Language:English
Published: MDPI AG 2022-05-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/10/10/1750
_version_ 1797498152824078336
author Tonglai Liu
Ronghai Luo
Longqin Xu
Dachun Feng
Liang Cao
Shuangyin Liu
Jianjun Guo
author_facet Tonglai Liu
Ronghai Luo
Longqin Xu
Dachun Feng
Liang Cao
Shuangyin Liu
Jianjun Guo
author_sort Tonglai Liu
collection DOAJ
description Recently, the attention mechanism combining spatial and channel information has been widely used in various deep convolutional neural networks (CNNs), proving its great potential in improving model performance. However, this usually uses 2D global pooling operations to compress spatial information or scaling methods to reduce the computational overhead in channel attention. These methods will result in severe information loss. Therefore, we propose a Spatial channel attention mechanism that captures cross-dimensional interaction, which does not involve dimensionality reduction and brings significant performance improvement with negligible computational overhead. The proposed attention mechanism can be seamlessly integrated into any convolutional neural network since it is a lightweight general module. Our method achieves a performance improvement of 2.08% on ResNet and 1.02% on MobileNetV2 in top-one error rate on the ImageNet dataset.
first_indexed 2024-03-10T03:29:23Z
format Article
id doaj.art-e11afd6c9907493d8a54e8530712249c
institution Directory Open Access Journal
issn 2227-7390
language English
last_indexed 2024-03-10T03:29:23Z
publishDate 2022-05-01
publisher MDPI AG
record_format Article
series Mathematics
spelling doaj.art-e11afd6c9907493d8a54e8530712249c2023-11-23T12:01:53ZengMDPI AGMathematics2227-73902022-05-011010175010.3390/math10101750Spatial Channel Attention for Deep Convolutional Neural NetworksTonglai Liu0Ronghai Luo1Longqin Xu2Dachun Feng3Liang Cao4Shuangyin Liu5Jianjun Guo6College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaSchool of Information and Communication, Guilin University of Electronic Technology, Guilin 541004, ChinaCollege of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaCollege of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaCollege of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaCollege of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaCollege of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, ChinaRecently, the attention mechanism combining spatial and channel information has been widely used in various deep convolutional neural networks (CNNs), proving its great potential in improving model performance. However, this usually uses 2D global pooling operations to compress spatial information or scaling methods to reduce the computational overhead in channel attention. These methods will result in severe information loss. Therefore, we propose a Spatial channel attention mechanism that captures cross-dimensional interaction, which does not involve dimensionality reduction and brings significant performance improvement with negligible computational overhead. The proposed attention mechanism can be seamlessly integrated into any convolutional neural network since it is a lightweight general module. Our method achieves a performance improvement of 2.08% on ResNet and 1.02% on MobileNetV2 in top-one error rate on the ImageNet dataset.https://www.mdpi.com/2227-7390/10/10/1750attention mechanismimage classificationdeep learningcross-dimensional interaction
spellingShingle Tonglai Liu
Ronghai Luo
Longqin Xu
Dachun Feng
Liang Cao
Shuangyin Liu
Jianjun Guo
Spatial Channel Attention for Deep Convolutional Neural Networks
Mathematics
attention mechanism
image classification
deep learning
cross-dimensional interaction
title Spatial Channel Attention for Deep Convolutional Neural Networks
title_full Spatial Channel Attention for Deep Convolutional Neural Networks
title_fullStr Spatial Channel Attention for Deep Convolutional Neural Networks
title_full_unstemmed Spatial Channel Attention for Deep Convolutional Neural Networks
title_short Spatial Channel Attention for Deep Convolutional Neural Networks
title_sort spatial channel attention for deep convolutional neural networks
topic attention mechanism
image classification
deep learning
cross-dimensional interaction
url https://www.mdpi.com/2227-7390/10/10/1750
work_keys_str_mv AT tonglailiu spatialchannelattentionfordeepconvolutionalneuralnetworks
AT ronghailuo spatialchannelattentionfordeepconvolutionalneuralnetworks
AT longqinxu spatialchannelattentionfordeepconvolutionalneuralnetworks
AT dachunfeng spatialchannelattentionfordeepconvolutionalneuralnetworks
AT liangcao spatialchannelattentionfordeepconvolutionalneuralnetworks
AT shuangyinliu spatialchannelattentionfordeepconvolutionalneuralnetworks
AT jianjunguo spatialchannelattentionfordeepconvolutionalneuralnetworks