Preserving Semantics in Textual Adversarial Attacks (2211.04205v2)
Abstract: The growth of hateful online content, or hate speech, has been associated with a global increase in violent crimes against minorities [23]. Harmful online content can be produced easily, automatically and anonymously. Even though, some form of auto-detection is already achieved through text classifiers in NLP, they can be fooled by adversarial attacks. To strengthen existing systems and stay ahead of attackers, we need better adversarial attacks. In this paper, we show that up to 70% of adversarial examples generated by adversarial attacks should be discarded because they do not preserve semantics. We address this core weakness and propose a new, fully supervised sentence embedding technique called Semantics-Preserving-Encoder (SPE). Our method outperforms existing sentence encoders used in adversarial attacks by achieving 1.2x - 5.1x better real attack success rate. We release our code as a plugin that can be used in any existing adversarial attack to improve its quality and speed up its execution.
- Generating natural language adversarial examples, 2018.
- Universal sentence encoder, 2018.
- Towards robustness against natural language word substitutions, 2021.
- Zachary Laub. Hate speech on social media: Global comparisons, 2019. Accessed: 2023-01-07.
- RoBERTa: A Robustly Optimized BERT Pretraining Approach, July 2019.
- BERTScore: Evaluating Text Generation with BERT, February 2020.
Collections
Sign up for free to add this paper to one or more collections.