@inproceedings{410ac1effb9d4534bd9eaf0b53914242,
title = "On the robustness of self-attentive models",
abstract = "This work examines the robustness of self-attentive neural networks against adversarial input perturbations. Specifically, we investigate the attention and feature extraction mechanisms of state-of-the-art recurrent neural networks and self-attentive architectures for sentiment analysis, entailment and machine translation under adversarial attacks. We also propose a novel attack algorithm for generating more natural adversarial examples that could mislead neural models but not humans. Experimental results show that, compared to recurrent neural models, self-attentive models are more robust against adversarial perturbation. In addition, we provide theoretical explanations for their superior robustness to support our claims.",
author = "Hsieh, \{Yu Lun\} and Minhao Cheng and Juan, \{Da Cheng\} and Wei Wei and Hsu, \{Wen Lian\} and Hsieh, \{Cho Jui\}",
note = "Publisher Copyright: {\textcopyright} 2019 Association for Computational Linguistics; 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019 ; Conference date: 28-07-2019 Through 02-08-2019",
year = "2020",
language = "English (US)",
series = "ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference",
publisher = "Association for Computational Linguistics (ACL)",
pages = "1520--1529",
booktitle = "ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference",
}