SOTAVerified

Learning to Discriminate Perturbations for Blocking Adversarial Attacks in Text Classification

2019-09-06IJCNLP 2019Code Available0· sign in to hype

Yichao Zhou, Jyun-Yu Jiang, Kai-Wei Chang, Wei Wang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Adversarial attacks against machine learning models have threatened various real-world applications such as spam filtering and sentiment analysis. In this paper, we propose a novel framework, learning to DIScriminate Perturbations (DISP), to identify and adjust malicious perturbations, thereby blocking adversarial attacks for text classification models. To identify adversarial attacks, a perturbation discriminator validates how likely a token in the text is perturbed and provides a set of potential perturbations. For each potential perturbation, an embedding estimator learns to restore the embedding of the original word based on the context and a replacement token is chosen based on approximate kNN search. DISP can block adversarial attacks for any NLP model without modifying the model structure or training procedure. Extensive experiments on two benchmark datasets demonstrate that DISP significantly outperforms baseline methods in blocking adversarial attacks for text classification. In addition, in-depth analysis shows the robustness of DISP across different situations.

Tasks

Reproductions