Prior convictions: Black-box adversarial attacks with bandits and priors

We study the problem of generating adversarial examples in a black-box setting in which only loss-oracle access to a model is available. We introduce a framework that conceptually unifies much of the existing work on black-box attacks, and we demonstrate that the current state-of-the-art methods are...

Full description

Bibliographic Details
Main Authors: Ilyas, Andrew., Engstrom, Logan G., Madry, Aleksander
Other Authors: MIT-IBM Watson AI Lab
Format: Article
Language:English
Published: arXiv 2021
Online Access:https://hdl.handle.net/1721.1/129721