A Closer Look into the Robustness of Neural Dependency Parsers Using Better Adversarial Examples
| Authors |
|
|---|---|
| Publication date | 2021 |
| Host editors |
|
| Book title | Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 |
| Book subtitle | Findings of ACL: ACL-IJCNLP 2021 : August 1-6, 2021 |
| ISBN (electronic) |
|
| Event | The Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-IJCNLP 2021) |
| Pages (from-to) | 2344-2354 |
| Number of pages | 11 |
| Publisher | Stroudsburg, PA: The Association for Computational Linguistics |
| Organisations |
|
| Abstract |
Previous work on adversarial attacks on dependency parsers has mostly focused on attack methods, as opposed to the quality of adversarial examples, which in previous work has been relatively low. To address this gap, we propose a method to generate high-quality adversarial examples with a higher number of candidate generators and stricter filters, and then verify their quality using automatic and human evaluations. We perform analysis with different parsing models and observe that: (i) injecting words not used in the training stage is an effective attack strategy; (ii) adversarial examples generated against a parser strongly depend on the parser model, the token embeddings, and even the specific instantiation of the model (i.e., a random seed). We use these insights to improve the robustness of English parsing models, relying on adversarial training and model ensembling. |
| Document type | Conference contribution |
| Note | With supplementary video |
| Language | English |
| Published at | https://doi.org/10.18653/v1/2021.findings-acl.207 |
| Other links | https://github.com/wangyuxuan93/depattacker https://www.scopus.com/pages/publications/85121708682 |
| Downloads |
2021.findings-acl.207
(Final published version)
|
| Supplementary materials | |
| Permalink to this page | |