Procedural noise adversarial examples for black-box attacks on deep neural networks

Autor: Emil Lupu, Luis Muñoz-González, Kenneth T. Co, Sixte de Maupeou
Rok vydání: 2019
Předmět:
FOS: Computer and information sciences
Technology
Computer Science - Cryptography and Security
Computer science
Machine Learning (stat.ML)
02 engineering and technology
Adversarial machine learning
Machine learning
computer.software_genre
adversarial machine learning
Adversarial system
cs.CR
Statistics - Machine Learning
Computer Science
Theory & Methods

0202 electrical engineering
electronic engineering
information engineering

black-box attacks
universal adversarial perturbations
Bayesian optimization
Science & Technology
Computer Science
Information Systems

business.industry
020206 networking & telecommunications
020207 software engineering
stat.ML
deep neural networks
Computer Science
Telecommunications
procedural noise
Artificial intelligence
business
computer
Cryptography and Security (cs.CR)
Zdroj: 26th ACM Conference on Computer and Communications Security
CCS
Popis: Deep Convolutional Networks (DCNs) have been shown to be vulnerable to adversarial examples---perturbed inputs specifically designed to produce intentional errors in the learning algorithms at test time. Existing input-agnostic adversarial perturbations exhibit interesting visual patterns that are currently unexplained. In this paper, we introduce a structured approach for generating Universal Adversarial Perturbations (UAPs) with procedural noise functions. Our approach unveils the systemic vulnerability of popular DCN models like Inception v3 and YOLO v3, with single noise patterns able to fool a model on up to 90% of the dataset. Procedural noise allows us to generate a distribution of UAPs with high universal evasion rates using only a few parameters. Additionally, we propose Bayesian optimization to efficiently learn procedural noise parameters to construct inexpensive untargeted black-box attacks. We demonstrate that it can achieve an average of less than 10 queries per successful attack, a 100-fold improvement on existing methods. We further motivate the use of input-agnostic defences to increase the stability of models to adversarial perturbations. The universality of our attacks suggests that DCN models may be sensitive to aggregations of low-level class-agnostic features. These findings give insight on the nature of some universal adversarial perturbations and how they could be generated in other applications.
Comment: 16 pages, 10 figures. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security (CCS '19)
Databáze: OpenAIRE