Towards certification of uncertainty calibration under adversarial attacks
<p>Since neural classifiers are known to be sensitive to adversarial perturbations that alter their accuracy, certification methods have been developed to provide provable guarantees on the insensitivity of their predictions to such perturbations. Furthermore, in safety-critical applications,...
Main Authors: | , , , , |
---|---|
Format: | Conference item |
Language: | English |
Published: |
Association for the Advancement of Artificial Intelligence
2024
|