Enforcing Interpretability in Time Series Transformers: A Concept Bottleneck Framework

Open Access
Authors
Publication date 08-10-2024
Edition v1
Number of pages 23
Publisher ArXiv
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
There has been a recent push of research on Transformer-based models for long-term time series forecasting, even though they are inherently difficult to interpret and explain. While there is a large body of work on interpretability methods for various domains and architectures, the interpretability of Transformer-based forecasting models remains largely unexplored. To address this gap, we develop a framework based on Concept Bottleneck Models to enforce interpretability of time series Transformers. We modify the training objective to encourage a model to develop representations similar to predefined interpretable concepts. In our experiments, we enforce similarity using Centered Kernel Alignment, and the predefined concepts include time features and an interpretable, autoregressive surrogate model (AR). We apply the framework to the Autoformer model, and present an in-depth analysis for a variety of benchmark tasks. We find that the model performance remains mostly unaffected, while the model shows much improved interpretability. Additionally, interpretable concepts become local, which makes the trained model easily intervenable. As a proof of concept, we demonstrate a successful intervention in the scenario of a time shift in the data, which eliminates the need to retrain.
Document type Preprint
Note v2 with title: Interpretability for Time Series Transformers using A Concept Bottleneck Framework.
Language English
Published at https://doi.org/10.48550/arXiv.2410.06070
Published at https://openreview.net/forum?id=A0mk2Wi68Y
Other links http://adsabs.harvard.edu/abs/2024arXiv241006070V
Downloads
2410.06070v1 (Final published version)
Permalink to this page
Back