Mathematical foundations of explainable AI and advances in bandit optimisation

Open Access
Authors
Supervisors
Cosupervisors
Award date 27-01-2026
ISBN
  • 9789465229386
Number of pages 242
Organisations
  • Faculty of Science (FNWI) - Korteweg-de Vries Institute for Mathematics (KdVI)
Abstract
This dissertation consists of two parts. In the first part it presents several works that expand the mathematical foundations explainable AI methods. In particular, it investigates attribution methods, counterfactual methods and concept-based models. Attribution methods indicate which input features are the most important to a particular model. What importance means is often ambiguous. We propose one interpretation in Chapter 2, where we interpret the attribution scores as a direction. The direction tells the user how to change their features to achieve a certain goal. We show that such methods are not robust with respect to the input: users with very similar attributes, might get drastically different explanations. In the following chapters 3 and 4, we zoom in on the counterfactual explanations. We demonstrate that following these explanations will change the underlying data distribution. We show that this can result in a decrease in accuracy of the model and even invalid the explanations themselves over time. In Chapter 5 we propose a method and new framework that can be used to develop sample-efficient concept-based models. By effectively leveraging the techniques used in Causal Representation Learning, we are able to be more data efficient. Finally, in the second part and the last chapter, we look at the bandit convex optimisation problem. We propose a new algorithm that is able to solve this problem, which has improved regret bounds compared to earlier algorithms, while being efficiently implementable.
Document type PhD thesis
Language English
Downloads
Permalink to this page
cover
Back