Bert is Robust! A Case Against Word Substitution-Based Adversarial Attacks

ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2023)

引用 0|浏览10
暂无评分
摘要
In this work, we investigate the robustness of BERT using four word substitution-based attacks. We combine a human evaluation of individual word substitutions and probabilistic analysis to show that most of the adversarial examples from the four studied attacks do not preserve enough semantics from the original examples, and can thus be easily recognized by human annotators. To further confirm that, we introduce an efficient adversarial defense consisting of a data augmentation step and a post-processing step. We show that many successful attacks can be defended using our defense method by including data similar to adversarial examples during training.
更多
查看译文
关键词
Language Models,Adversarial Attacks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要