On the hardness of robust classification

It is becoming increasingly important to understand the vulnerability of machine learning models to adversarial attacks. In this paper we study the feasibility of adversarially robust learning from the perspective of computational learning theory, considering both sample and computational complexity...

Ausführliche Beschreibung

Bibliographische Detailangaben
Hauptverfasser: Gourdeau, P, Kanade, V, Kwiatkowska, M, Worrell, J
Format: Journal article
Sprache:English
Veröffentlicht: Journal of Machine Learning Research 2021
_version_ 1826299627767332864
author Gourdeau, P
Kanade, V
Kwiatkowska, M
Worrell, J
author_facet Gourdeau, P
Kanade, V
Kwiatkowska, M
Worrell, J
author_sort Gourdeau, P
collection OXFORD
description It is becoming increasingly important to understand the vulnerability of machine learning models to adversarial attacks. In this paper we study the feasibility of adversarially robust learning from the perspective of computational learning theory, considering both sample and computational complexity. In particular, our definition of robust learnability requires polynomial sample complexity. We start with two negative results. We show that no non-trivial concept class can be robustly learned in the distribution-free setting against an adversary who can perturb just a single input bit. We show, moreover, that the class of monotone conjunctions cannot be robustly learned under the uniform distribution against an adversary who can perturb ω(logn) input bits. However, we also show that if the adversary is restricted to perturbing O(logn) bits, then one can robustly learn the class of 1-decision lists (which subsumes monotone conjunctions) with respect to the class of log-Lipschitz distributions. We then extend this result to show learnability of 2-decision lists and monotone k-decision lists in the same distributional and adversarial setting. Finally, we provide a simple proof of the computational hardness of robust learning on the boolean hypercube. Unlike previous results of this nature, our result does not rely on a more restricted model of learning, such as the statistical query model, nor on any hardness assumption other than the existence of an (average-case) hard learning problem in the PAC framework; this allows us to have a clean proof of the reduction, and the assumption is no stronger than assumptions that are used to build cryptographic primitives.
first_indexed 2024-03-07T05:04:49Z
format Journal article
id oxford-uuid:d98e9853-2fe3-45d4-a2d1-f6ed85e6df01
institution University of Oxford
language English
last_indexed 2024-03-07T05:04:49Z
publishDate 2021
publisher Journal of Machine Learning Research
record_format dspace
spelling oxford-uuid:d98e9853-2fe3-45d4-a2d1-f6ed85e6df012022-03-27T08:56:45ZOn the hardness of robust classificationJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:d98e9853-2fe3-45d4-a2d1-f6ed85e6df01EnglishSymplectic ElementsJournal of Machine Learning Research2021Gourdeau, PKanade, VKwiatkowska, MWorrell, JIt is becoming increasingly important to understand the vulnerability of machine learning models to adversarial attacks. In this paper we study the feasibility of adversarially robust learning from the perspective of computational learning theory, considering both sample and computational complexity. In particular, our definition of robust learnability requires polynomial sample complexity. We start with two negative results. We show that no non-trivial concept class can be robustly learned in the distribution-free setting against an adversary who can perturb just a single input bit. We show, moreover, that the class of monotone conjunctions cannot be robustly learned under the uniform distribution against an adversary who can perturb ω(logn) input bits. However, we also show that if the adversary is restricted to perturbing O(logn) bits, then one can robustly learn the class of 1-decision lists (which subsumes monotone conjunctions) with respect to the class of log-Lipschitz distributions. We then extend this result to show learnability of 2-decision lists and monotone k-decision lists in the same distributional and adversarial setting. Finally, we provide a simple proof of the computational hardness of robust learning on the boolean hypercube. Unlike previous results of this nature, our result does not rely on a more restricted model of learning, such as the statistical query model, nor on any hardness assumption other than the existence of an (average-case) hard learning problem in the PAC framework; this allows us to have a clean proof of the reduction, and the assumption is no stronger than assumptions that are used to build cryptographic primitives.
spellingShingle Gourdeau, P
Kanade, V
Kwiatkowska, M
Worrell, J
On the hardness of robust classification
title On the hardness of robust classification
title_full On the hardness of robust classification
title_fullStr On the hardness of robust classification
title_full_unstemmed On the hardness of robust classification
title_short On the hardness of robust classification
title_sort on the hardness of robust classification
work_keys_str_mv AT gourdeaup onthehardnessofrobustclassification
AT kanadev onthehardnessofrobustclassification
AT kwiatkowskam onthehardnessofrobustclassification
AT worrellj onthehardnessofrobustclassification