Adversarial Examples and Distribution Shift: A Representations Perspective

Adversarial attacks cause machine learning models to produce wrong predictions by minimally perturbing their input. In this thesis, we take a step towards understanding how these perturbations affect the intermediate data representations of the model. Specifically, we compare standard and adversaria...

Full description

Bibliographic Details
Main Author: Nadhamuni, Kaveri
Other Authors: Madry, Aleksander
Format: Thesis
Published: Massachusetts Institute of Technology 2022
Online Access:https://hdl.handle.net/1721.1/138945