A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity
Research on visual encoding models for functional magnetic resonance imaging derived from deep neural networks, especially CNN (e.g., VGG16), has been developed. However, CNNs typically use smaller kernel sizes (e.g., 3 × 3) for feature extraction in visual encoding models. Although the receptive fi...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-11-01
|
Series: | Brain Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3425/12/12/1633 |
_version_ | 1797461184600866816 |
---|---|
author | Shuxiao Ma Linyuan Wang Panpan Chen Ruoxi Qin Libin Hou Bin Yan |
author_facet | Shuxiao Ma Linyuan Wang Panpan Chen Ruoxi Qin Libin Hou Bin Yan |
author_sort | Shuxiao Ma |
collection | DOAJ |
description | Research on visual encoding models for functional magnetic resonance imaging derived from deep neural networks, especially CNN (e.g., VGG16), has been developed. However, CNNs typically use smaller kernel sizes (e.g., 3 × 3) for feature extraction in visual encoding models. Although the receptive field size of CNN can be enlarged by increasing the network depth or subsampling, it is limited by the small size of the convolution kernel, leading to an insufficient receptive field size. In biological research, the size of the neuronal population receptive field of high-level visual encoding regions is usually three to four times that of low-level visual encoding regions. Thus, CNNs with a larger receptive field size align with the biological findings. The RepLKNet model directly expands the convolution kernel size to obtain a larger-scale receptive field. Therefore, this paper proposes a mixed model to replace CNN for feature extraction in visual encoding models. The proposed model mixes RepLKNet and VGG so that the mixed model has a receptive field of different sizes to extract more feature information from the image. The experimental results indicate that the mixed model achieves better encoding performance in multiple regions of the visual cortex than the traditional convolutional model. Also, a larger-scale receptive field should be considered in building visual encoding models so that the convolution network can play a more significant role in visual representations. |
first_indexed | 2024-03-09T17:16:46Z |
format | Article |
id | doaj.art-f1a202aab5f74d15b38636c8116b516b |
institution | Directory Open Access Journal |
issn | 2076-3425 |
language | English |
last_indexed | 2024-03-09T17:16:46Z |
publishDate | 2022-11-01 |
publisher | MDPI AG |
record_format | Article |
series | Brain Sciences |
spelling | doaj.art-f1a202aab5f74d15b38636c8116b516b2023-11-24T13:39:04ZengMDPI AGBrain Sciences2076-34252022-11-011212163310.3390/brainsci12121633A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain ActivityShuxiao Ma0Linyuan Wang1Panpan Chen2Ruoxi Qin3Libin Hou4Bin Yan5Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaHenan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaHenan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaHenan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaHenan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaHenan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information, Engineering University, Zhengzhou 450001, ChinaResearch on visual encoding models for functional magnetic resonance imaging derived from deep neural networks, especially CNN (e.g., VGG16), has been developed. However, CNNs typically use smaller kernel sizes (e.g., 3 × 3) for feature extraction in visual encoding models. Although the receptive field size of CNN can be enlarged by increasing the network depth or subsampling, it is limited by the small size of the convolution kernel, leading to an insufficient receptive field size. In biological research, the size of the neuronal population receptive field of high-level visual encoding regions is usually three to four times that of low-level visual encoding regions. Thus, CNNs with a larger receptive field size align with the biological findings. The RepLKNet model directly expands the convolution kernel size to obtain a larger-scale receptive field. Therefore, this paper proposes a mixed model to replace CNN for feature extraction in visual encoding models. The proposed model mixes RepLKNet and VGG so that the mixed model has a receptive field of different sizes to extract more feature information from the image. The experimental results indicate that the mixed model achieves better encoding performance in multiple regions of the visual cortex than the traditional convolutional model. Also, a larger-scale receptive field should be considered in building visual encoding models so that the convolution network can play a more significant role in visual representations.https://www.mdpi.com/2076-3425/12/12/1633visual encoding modelsdeep neural networksreceptive fielda large convolution kernelRepLKNetfMRI |
spellingShingle | Shuxiao Ma Linyuan Wang Panpan Chen Ruoxi Qin Libin Hou Bin Yan A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity Brain Sciences visual encoding models deep neural networks receptive field a large convolution kernel RepLKNet fMRI |
title | A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity |
title_full | A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity |
title_fullStr | A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity |
title_full_unstemmed | A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity |
title_short | A Mixed Visual Encoding Model Based on the Larger-Scale Receptive Field for Human Brain Activity |
title_sort | mixed visual encoding model based on the larger scale receptive field for human brain activity |
topic | visual encoding models deep neural networks receptive field a large convolution kernel RepLKNet fMRI |
url | https://www.mdpi.com/2076-3425/12/12/1633 |
work_keys_str_mv | AT shuxiaoma amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT linyuanwang amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT panpanchen amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT ruoxiqin amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT libinhou amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT binyan amixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT shuxiaoma mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT linyuanwang mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT panpanchen mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT ruoxiqin mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT libinhou mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity AT binyan mixedvisualencodingmodelbasedonthelargerscalereceptivefieldforhumanbrainactivity |