Autor: |
Gaoming Yang, Mingwei Li, Xianjing Fang, Ji Zhang, Xingzhu Liang |
Jazyk: |
angličtina |
Rok vydání: |
2021 |
Předmět: |
|
Zdroj: |
PeerJ Computer Science, Vol 7, p e702 (2021) |
Druh dokumentu: |
article |
ISSN: |
2376-5992 |
DOI: |
10.7717/peerj-cs.702 |
Popis: |
Adversarial examples are regarded as a security threat to deep learning models, and there are many ways to generate them. However, most existing methods require the query authority of the target during their work. In a more practical situation, the attacker will be easily detected because of too many queries, and this problem is especially obvious under the black-box setting. To solve the problem, we propose the Attack Without a Target Model (AWTM). Our algorithm does not specify any target model in generating adversarial examples, so it does not need to query the target. Experimental results show that it achieved a maximum attack success rate of 81.78% in the MNIST data set and 87.99% in the CIFAR-10 data set. In addition, it has a low time cost because it is a GAN-based method. |
Databáze: |
Directory of Open Access Journals |
Externí odkaz: |
|