A Universal Adversarial Policy for Text Classifiers

Gallil Maimon, Lior Rokach

Research output: Contribution to journalArticlepeer-review

40 Downloads (Pure)


Discovering the existence of universal adversarial perturbations had large theoretical and practical impacts on the field of adversarial learning. In the text domain, most universal studies focused on adversarial prefixes which are added to all texts. However, unlike the vision domain, adding the same perturbation to different inputs results in noticeably unnatural inputs. Therefore, we introduce a new universal adversarial setup - a universal adversarial policy, which has many advantages of other universal attacks but also results in valid texts - thus making it relevant in practice. We achieve this by learning a single search policy over a predefined set of semantics preserving text alterations, on many texts. This formulation is universal in that the policy is successful in finding adversarial examples on new texts efficiently. Our approach uses text perturbations which were extensively shown to produce natural attacks in the non-universal setup (specific synonym replacements). We suggest a strong baseline approach for this formulation which uses reinforcement learning. It's ability to generalise (from as few as 500 training texts) shows that universal adversarial patterns exist in the text domain as well.
Original languageEnglish
Pages (from-to)282-291
JournalNeural Networks
StatePublished - 1 Jun 2022


  • Computer Science - Machine Learning
  • Computer Science - Computation and Language
  • Computer Science - Cryptography and Security


Dive into the research topics of 'A Universal Adversarial Policy for Text Classifiers'. Together they form a unique fingerprint.

Cite this