TY - JOUR
T1 - Efficiently generating sentence-level textual adversarial examples with Seq2seq Stacked Auto-Encoder
AU - Li, Ang
AU - Zhang, Fangyuan
AU - Li, Shuangjiao
AU - Chen, Tianhua
AU - Su, Pan
AU - Wang, Hongtao
N1 - Funding Information:
This work was supported in part by the National Natural Science Foundation of China (Grant No. 61802124 ), and the Fundamental Research Funds for the Central Universities, China (Grant No. 2019MS126 ).
Publisher Copyright:
© 2022 Elsevier Ltd
PY - 2023/3/1
Y1 - 2023/3/1
N2 - In spite deep learning has advanced numerous successes, recent research has shown increasing concern on its vulnerability over adversarial attacks. In Natural Language Processing, crafting high-quality adversarial text examples is much more challenging due to the discrete nature of texts. Recent studies perform transformations on characters or words, which are generally formulated as combinatorial optimization problems. However, these approaches suffer from inefficiency due to the high dimensional search space. To address this issue, in this paper, we propose an end-to-end Seq2seq Stacked Auto-Encoder (SSAE) neural network, which generates adversarial text examples efficiently via direct network inference. SSAE has two salient features. The outer auto-encoder preserves syntactic and semantic information to the original examples. The inner auto-encoder projects sentence embedding into a high-level semantic representation, on which constrained perturbations are superimposed to increase adversarial ability. Experimental results suggest that SSAE has a higher attack success rate than existing word-level attack methods, and is 100x to 700x faster at attack speed on IMDB dataset. We further find out that the adversarial examples generated by SSAE have strong transferability to attack different victim models.
AB - In spite deep learning has advanced numerous successes, recent research has shown increasing concern on its vulnerability over adversarial attacks. In Natural Language Processing, crafting high-quality adversarial text examples is much more challenging due to the discrete nature of texts. Recent studies perform transformations on characters or words, which are generally formulated as combinatorial optimization problems. However, these approaches suffer from inefficiency due to the high dimensional search space. To address this issue, in this paper, we propose an end-to-end Seq2seq Stacked Auto-Encoder (SSAE) neural network, which generates adversarial text examples efficiently via direct network inference. SSAE has two salient features. The outer auto-encoder preserves syntactic and semantic information to the original examples. The inner auto-encoder projects sentence embedding into a high-level semantic representation, on which constrained perturbations are superimposed to increase adversarial ability. Experimental results suggest that SSAE has a higher attack success rate than existing word-level attack methods, and is 100x to 700x faster at attack speed on IMDB dataset. We further find out that the adversarial examples generated by SSAE have strong transferability to attack different victim models.
KW - Sentence-level attack
KW - Textual adversarial examples
KW - Deep neural network
KW - Stacked auto-encoder
UR - http://www.scopus.com/inward/record.url?scp=85141925714&partnerID=8YFLogxK
U2 - 10.1016/j.eswa.2022.119170
DO - 10.1016/j.eswa.2022.119170
M3 - Article
VL - 213
JO - Expert Systems with Applications
JF - Expert Systems with Applications
SN - 0957-4174
IS - Part C
M1 - 119170
ER -