@inproceedings{45e8b846b9f346e382b1d1109f44a7dd,
title = "Generation of low distortion adversarial attacks via convex programming",
abstract = "As deep neural networks (DNNs) achieve extraordinary performance in a wide range of tasks, testing their robustness under adversarial attacks becomes paramount. Adversarial attacks, also known as adversarial examples, are used to measure the robustness of DNNs and are generated by incorporating imperceptible perturbations into the input data with the intention of altering a DNN's classification. In prior work in this area, most of the proposed optimization based methods employ gradient descent to find adversarial examples. In this paper, we present an innovative method which generates adversarial examples via convex programming. Our experiment results demonstrate that we can generate adversarial examples with lower distortion and higher transferability than the C&W attack, which is the current state-of-the-art adversarial attack method for DNNs. We achieve 100% attack success rate on both the original undefended models and the adversarially-trained models. Our distortions of the L-inf attack are respectively 31% and 18% lower than the C&W attack for the best case and average case on the CIFAR-10 data set.",
keywords = "Adversarial attack, Convex programming, Deep neural networks",
author = "Tianyun Zhang and Sijia Liu and Yanzhi Wang and Makan Fardad",
year = "2019",
month = nov,
doi = "10.1109/ICDM.2019.00195",
language = "English (US)",
series = "Proceedings - IEEE International Conference on Data Mining, ICDM",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "1486--1491",
editor = "Jianyong Wang and Kyuseok Shim and Xindong Wu",
booktitle = "Proceedings - 19th IEEE International Conference on Data Mining, ICDM 2019",
note = "19th IEEE International Conference on Data Mining, ICDM 2019 ; Conference date: 08-11-2019 Through 11-11-2019",
}