Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks

Despite the enormous success of artificial neural networks (ANNs) in many disciplines, the characterization of their computations and the origin of key properties such as generalization and robustness remain open questions. Recent literature suggests that robust networks with good generalization pro...

Full description

Bibliographic Details
Main Authors: Nikos Karantzas, Emma Besier, Josue Ortega Caro, Xaq Pitkow, Andreas S. Tolias, Ankit B. Patel, Fabio Anselmi
Format: Article
Language:English
Published: Frontiers Media S.A. 2022-07-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2022.890016/full
_version_ 1818108390855933952
author Nikos Karantzas
Nikos Karantzas
Emma Besier
Emma Besier
Josue Ortega Caro
Josue Ortega Caro
Xaq Pitkow
Xaq Pitkow
Xaq Pitkow
Andreas S. Tolias
Andreas S. Tolias
Andreas S. Tolias
Ankit B. Patel
Ankit B. Patel
Ankit B. Patel
Fabio Anselmi
Fabio Anselmi
author_facet Nikos Karantzas
Nikos Karantzas
Emma Besier
Emma Besier
Josue Ortega Caro
Josue Ortega Caro
Xaq Pitkow
Xaq Pitkow
Xaq Pitkow
Andreas S. Tolias
Andreas S. Tolias
Andreas S. Tolias
Ankit B. Patel
Ankit B. Patel
Ankit B. Patel
Fabio Anselmi
Fabio Anselmi
author_sort Nikos Karantzas
collection DOAJ
description Despite the enormous success of artificial neural networks (ANNs) in many disciplines, the characterization of their computations and the origin of key properties such as generalization and robustness remain open questions. Recent literature suggests that robust networks with good generalization properties tend to be biased toward processing low frequencies in images. To explore the frequency bias hypothesis further, we develop an algorithm that allows us to learn modulatory masks highlighting the essential input frequencies needed for preserving a trained network's performance. We achieve this by imposing invariance in the loss with respect to such modulations in the input frequencies. We first use our method to test the low-frequency preference hypothesis of adversarially trained or data-augmented networks. Our results suggest that adversarially robust networks indeed exhibit a low-frequency bias but we find this bias is also dependent on directions in frequency space. However, this is not necessarily true for other types of data augmentation. Our results also indicate that the essential frequencies in question are effectively the ones used to achieve generalization in the first place. Surprisingly, images seen through these modulatory masks are not recognizable and resemble texture-like patterns.
first_indexed 2024-12-11T02:14:36Z
format Article
id doaj.art-7ea7bb52abb44d12a7344e304625acda
institution Directory Open Access Journal
issn 2624-8212
language English
last_indexed 2024-12-11T02:14:36Z
publishDate 2022-07-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Artificial Intelligence
spelling doaj.art-7ea7bb52abb44d12a7344e304625acda2022-12-22T01:24:12ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122022-07-01510.3389/frai.2022.890016890016Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier MasksNikos Karantzas0Nikos Karantzas1Emma Besier2Emma Besier3Josue Ortega Caro4Josue Ortega Caro5Xaq Pitkow6Xaq Pitkow7Xaq Pitkow8Andreas S. Tolias9Andreas S. Tolias10Andreas S. Tolias11Ankit B. Patel12Ankit B. Patel13Ankit B. Patel14Fabio Anselmi15Fabio Anselmi16Department of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Electrical and Computer Engineering, Rice University, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Electrical and Computer Engineering, Rice University, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDepartment of Electrical and Computer Engineering, Rice University, Houston, TX, United StatesDepartment of Neuroscience, Baylor College of Medicine, Houston, TX, United StatesCenter for Neuroscience and Artificial Intelligence, Baylor College of Medicine, Houston, TX, United StatesDespite the enormous success of artificial neural networks (ANNs) in many disciplines, the characterization of their computations and the origin of key properties such as generalization and robustness remain open questions. Recent literature suggests that robust networks with good generalization properties tend to be biased toward processing low frequencies in images. To explore the frequency bias hypothesis further, we develop an algorithm that allows us to learn modulatory masks highlighting the essential input frequencies needed for preserving a trained network's performance. We achieve this by imposing invariance in the loss with respect to such modulations in the input frequencies. We first use our method to test the low-frequency preference hypothesis of adversarially trained or data-augmented networks. Our results suggest that adversarially robust networks indeed exhibit a low-frequency bias but we find this bias is also dependent on directions in frequency space. However, this is not necessarily true for other types of data augmentation. Our results also indicate that the essential frequencies in question are effectively the ones used to achieve generalization in the first place. Surprisingly, images seen through these modulatory masks are not recognizable and resemble texture-like patterns.https://www.frontiersin.org/articles/10.3389/frai.2022.890016/fullFourier analysissymmetryrobustnessgeneralizationneural networksdata augmentation
spellingShingle Nikos Karantzas
Nikos Karantzas
Emma Besier
Emma Besier
Josue Ortega Caro
Josue Ortega Caro
Xaq Pitkow
Xaq Pitkow
Xaq Pitkow
Andreas S. Tolias
Andreas S. Tolias
Andreas S. Tolias
Ankit B. Patel
Ankit B. Patel
Ankit B. Patel
Fabio Anselmi
Fabio Anselmi
Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
Frontiers in Artificial Intelligence
Fourier analysis
symmetry
robustness
generalization
neural networks
data augmentation
title Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
title_full Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
title_fullStr Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
title_full_unstemmed Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
title_short Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
title_sort understanding robustness and generalization of artificial neural networks through fourier masks
topic Fourier analysis
symmetry
robustness
generalization
neural networks
data augmentation
url https://www.frontiersin.org/articles/10.3389/frai.2022.890016/full
work_keys_str_mv AT nikoskarantzas understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT nikoskarantzas understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT emmabesier understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT emmabesier understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT josueortegacaro understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT josueortegacaro understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT xaqpitkow understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT xaqpitkow understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT xaqpitkow understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT andreasstolias understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT andreasstolias understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT andreasstolias understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT ankitbpatel understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT ankitbpatel understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT ankitbpatel understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT fabioanselmi understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks
AT fabioanselmi understandingrobustnessandgeneralizationofartificialneuralnetworksthroughfouriermasks