On-manifold projected gradient descent

This study provides a computable, direct, and mathematically rigorous approximation to the differential geometry of class manifolds for high-dimensional data, along with non-linear projections from input space onto these class manifolds. The tools are applied to the setting of neural network image c...

Full description

Bibliographic Details
Main Authors: Aaron Mahler, Tyrus Berry, Tom Stephens, Harbir Antil, Michael Merritt, Jeanie Schreiber, Ioannis Kevrekidis
Format: Article
Language:English
Published: Frontiers Media S.A. 2024-02-01
Series:Frontiers in Computer Science
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fcomp.2024.1274181/full
_version_ 1797311682727378944
author Aaron Mahler
Tyrus Berry
Tom Stephens
Harbir Antil
Michael Merritt
Jeanie Schreiber
Ioannis Kevrekidis
author_facet Aaron Mahler
Tyrus Berry
Tom Stephens
Harbir Antil
Michael Merritt
Jeanie Schreiber
Ioannis Kevrekidis
author_sort Aaron Mahler
collection DOAJ
description This study provides a computable, direct, and mathematically rigorous approximation to the differential geometry of class manifolds for high-dimensional data, along with non-linear projections from input space onto these class manifolds. The tools are applied to the setting of neural network image classifiers, where we generate novel, on-manifold data samples and implement a projected gradient descent algorithm for on-manifold adversarial training. The susceptibility of neural networks (NNs) to adversarial attack highlights the brittle nature of NN decision boundaries in input space. Introducing adversarial examples during training has been shown to reduce the susceptibility of NNs to adversarial attack; however, it has also been shown to reduce the accuracy of the classifier if the examples are not valid examples for that class. Realistic “on-manifold” examples have been previously generated from class manifolds in the latent space of an autoencoder. Our study explores these phenomena in a geometric and computational setting that is much closer to the raw, high-dimensional input space than what can be provided by VAE or other black box dimensionality reductions. We employ conformally invariant diffusion maps (CIDM) to approximate class manifolds in diffusion coordinates and develop the Nyström projection to project novel points onto class manifolds in this setting. On top of the manifold approximation, we leverage the spectral exterior calculus (SEC) to determine geometric quantities such as tangent vectors of the manifold. We use these tools to obtain adversarial examples that reside on a class manifold, yet fool a classifier. These misclassifications then become explainable in terms of human-understandable manipulations within the data, by expressing the on-manifold adversary in the semantic basis on the manifold.
first_indexed 2024-03-08T02:03:14Z
format Article
id doaj.art-e48a2ae32da744189793198b22faf721
institution Directory Open Access Journal
issn 2624-9898
language English
last_indexed 2024-03-08T02:03:14Z
publishDate 2024-02-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Computer Science
spelling doaj.art-e48a2ae32da744189793198b22faf7212024-02-14T04:29:47ZengFrontiers Media S.A.Frontiers in Computer Science2624-98982024-02-01610.3389/fcomp.2024.12741811274181On-manifold projected gradient descentAaron Mahler0Tyrus Berry1Tom Stephens2Harbir Antil3Michael Merritt4Jeanie Schreiber5Ioannis Kevrekidis6Teledyne Scientific & Imaging, LLC, Durham, NC, United StatesCenter for Mathematics and Artificial Intelligence, George Mason University, Fairfax, VA, United StatesTeledyne Scientific & Imaging, LLC, Durham, NC, United StatesCenter for Mathematics and Artificial Intelligence, George Mason University, Fairfax, VA, United StatesTeledyne Scientific & Imaging, LLC, Durham, NC, United StatesCenter for Mathematics and Artificial Intelligence, George Mason University, Fairfax, VA, United StatesDepartments of Chemical and Biomolecular Engineering and Applied Mathematics and Statistics, Johns Hopkins University, Baltimore, MD, United StatesThis study provides a computable, direct, and mathematically rigorous approximation to the differential geometry of class manifolds for high-dimensional data, along with non-linear projections from input space onto these class manifolds. The tools are applied to the setting of neural network image classifiers, where we generate novel, on-manifold data samples and implement a projected gradient descent algorithm for on-manifold adversarial training. The susceptibility of neural networks (NNs) to adversarial attack highlights the brittle nature of NN decision boundaries in input space. Introducing adversarial examples during training has been shown to reduce the susceptibility of NNs to adversarial attack; however, it has also been shown to reduce the accuracy of the classifier if the examples are not valid examples for that class. Realistic “on-manifold” examples have been previously generated from class manifolds in the latent space of an autoencoder. Our study explores these phenomena in a geometric and computational setting that is much closer to the raw, high-dimensional input space than what can be provided by VAE or other black box dimensionality reductions. We employ conformally invariant diffusion maps (CIDM) to approximate class manifolds in diffusion coordinates and develop the Nyström projection to project novel points onto class manifolds in this setting. On top of the manifold approximation, we leverage the spectral exterior calculus (SEC) to determine geometric quantities such as tangent vectors of the manifold. We use these tools to obtain adversarial examples that reside on a class manifold, yet fool a classifier. These misclassifications then become explainable in terms of human-understandable manipulations within the data, by expressing the on-manifold adversary in the semantic basis on the manifold.https://www.frontiersin.org/articles/10.3389/fcomp.2024.1274181/fulldiffusion mapskernel methodsmanifold learningNyström approximationadversarial attackimage classification
spellingShingle Aaron Mahler
Tyrus Berry
Tom Stephens
Harbir Antil
Michael Merritt
Jeanie Schreiber
Ioannis Kevrekidis
On-manifold projected gradient descent
Frontiers in Computer Science
diffusion maps
kernel methods
manifold learning
Nyström approximation
adversarial attack
image classification
title On-manifold projected gradient descent
title_full On-manifold projected gradient descent
title_fullStr On-manifold projected gradient descent
title_full_unstemmed On-manifold projected gradient descent
title_short On-manifold projected gradient descent
title_sort on manifold projected gradient descent
topic diffusion maps
kernel methods
manifold learning
Nyström approximation
adversarial attack
image classification
url https://www.frontiersin.org/articles/10.3389/fcomp.2024.1274181/full
work_keys_str_mv AT aaronmahler onmanifoldprojectedgradientdescent
AT tyrusberry onmanifoldprojectedgradientdescent
AT tomstephens onmanifoldprojectedgradientdescent
AT harbirantil onmanifoldprojectedgradientdescent
AT michaelmerritt onmanifoldprojectedgradientdescent
AT jeanieschreiber onmanifoldprojectedgradientdescent
AT ioanniskevrekidis onmanifoldprojectedgradientdescent