Asymmetries in Adversarial Settings


Autoria(s): Marro, Samuele
Contribuinte(s)

Lombardi, Michele

Data(s)

13/07/2022

Resumo

Even without formal guarantees of their effectiveness, adversarial attacks against Machine Learning models frequently fool new defenses. We identify six key asymmetries that contribute to this phenomenon and formulate four guidelines to build future-proof defenses by preventing such asymmetries. We also prove that attacking a classifier is NP-complete, while defending from such attacks is Sigma_2^P-complete. We then introduce Counter-Attack (CA), an asymmetry-free metadefense that determines whether a model is robust on a given input by estimating its distance from the decision boundary. Under specific assumptions CA can provide theoretical detection guarantees. Additionally, we prove that while CA is NP-complete, fooling CA is Sigma_2^P-complete. Even when using heuristic relaxations, we show that our method can reliably identify non-robust points. As part of our experimental evaluation, we introduce UG100, a new dataset obtained by applying a provably optimal attack to six limited-scale networks (three for MNIST and three for CIFAR10), each trained in three different manners.

Formato

application/pdf

Identificador

http://amslaurea.unibo.it/26198/1/Asymmetries%20in%20Adversarial%20Settings%20-%20Samuele%20Marro.pdf

Marro, Samuele (2022) Asymmetries in Adversarial Settings. [Laurea], Università di Bologna, Corso di Studio in Informatica [L-DM270] <http://amslaurea.unibo.it/view/cds/CDS8009/>

Idioma(s)

en

Publicador

Alma Mater Studiorum - Università di Bologna

Relação

http://amslaurea.unibo.it/26198/

Direitos

cc_by_sa4

Palavras-Chave #machine learning,adversarial attacks,robustness,neural networks,mixed integer programming,adversarial asymmetries,metadefense,dataset #Informatica [L-DM270]
Tipo

PeerReviewed

info:eu-repo/semantics/bachelorThesis