Asymmetries in Adversarial Settings
Contribuinte(s) |
Lombardi, Michele |
---|---|
Data(s) |
13/07/2022
|
Resumo |
Even without formal guarantees of their effectiveness, adversarial attacks against Machine Learning models frequently fool new defenses. We identify six key asymmetries that contribute to this phenomenon and formulate four guidelines to build future-proof defenses by preventing such asymmetries. We also prove that attacking a classifier is NP-complete, while defending from such attacks is Sigma_2^P-complete. We then introduce Counter-Attack (CA), an asymmetry-free metadefense that determines whether a model is robust on a given input by estimating its distance from the decision boundary. Under specific assumptions CA can provide theoretical detection guarantees. Additionally, we prove that while CA is NP-complete, fooling CA is Sigma_2^P-complete. Even when using heuristic relaxations, we show that our method can reliably identify non-robust points. As part of our experimental evaluation, we introduce UG100, a new dataset obtained by applying a provably optimal attack to six limited-scale networks (three for MNIST and three for CIFAR10), each trained in three different manners. |
Formato |
application/pdf |
Identificador |
Marro, Samuele (2022) Asymmetries in Adversarial Settings. [Laurea], Università di Bologna, Corso di Studio in Informatica [L-DM270] <http://amslaurea.unibo.it/view/cds/CDS8009/> |
Idioma(s) |
en |
Publicador |
Alma Mater Studiorum - Università di Bologna |
Relação |
http://amslaurea.unibo.it/26198/ |
Direitos |
cc_by_sa4 |
Palavras-Chave | #machine learning,adversarial attacks,robustness,neural networks,mixed integer programming,adversarial asymmetries,metadefense,dataset #Informatica [L-DM270] |
Tipo |
PeerReviewed info:eu-repo/semantics/bachelorThesis |