Can Algorithms be Explained Without Compromising Efficiency? The Benefits of Detection and Imitation in Strategic Classification Extended abstract

Open Access
Authors
Publication date 2022
Book title AAMAS '22
Book subtitle Proceedings of the 21st International Conference on Autonomous Agents and Multiagent Systems : May 9-13, 2022, virtual event, New Zealand
ISBN
  • 9781713854333
ISBN (electronic)
  • 9781450392136
Event 21st International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2022
Volume | Issue number 3
Pages (from-to) 1536-1538
Number of pages 3
Publisher Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract

Given the ubiquity of AI-based decisions that affect individuals' lives, providing transparent explanations about algorithms is ethically sound and often legally mandatory. How do individuals strategically adapt following explanations? What are the consequences of adaptation for algorithmic accuracy? We simulate the interplay between explanations shared by an Institution (e.g. a bank) and the dynamics of strategic adaptation by Individuals reacting to such feedback. Resorting to an agent-based approach, our model scrutinizes the role of: i) transparency in explanations, ii) detection capacity and iii) behavior imitation. We find that the risks of transparent explanations are alleviated if effective methods to detect faking behaviors are in place. Furthermore, we observe that social learning and imitation - as often observed across societies - is likely to alleviate the impacts of (malicious) adaptation.

Document type Conference contribution
Language English
Published at https://dl.acm.org/doi/10.5555/3535850.3536026 https://www.ifaamas.org/Proceedings/aamas2022/pdfs/p1536.pdf
Other links https://www.scopus.com/pages/publications/85134330289
Downloads
p1536-1 (Final published version)
Permalink to this page
Back