An Information-theoretic Approach to Distribution Shifts

Open Access
Authors
Publication date 2022
Host editors
  • M. Ranzato
  • A. Beygelzimer
  • Y. Dauphin
  • P.S. Liang
  • J. Wortman Vaughan
Book title 35th Conference on Neural Information Processing Systems (NeurIPS 2021)
Book subtitle online, 6-14 December 2021
ISBN
  • 9781713845393
Series Advances in Neural Information Processing Systems
Event NeurIPS 2021
Volume | Issue number 21
Pages (from-to) 17628-17641
Publisher San Diego, CA: Neural Information Processing Systems Foundation
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Safely deploying machine learning models to the real world is often a challenging process. Models trained with data obtained from a specific geographic location tend to fail when queried with data obtained elsewhere, agents trained in a simulation can struggle to adapt when deployed in the real world or novel environments, and neural networks that are fit to a subset of the population might carry some selection bias into their decision process. In this work, we describe the problem of data shift from a novel information-theoretic perspective by (i) identifying and describing the different sources of error, (ii) comparing some of the most promising objectives explored in the recent domain generalization and fair classification literature. From our theoretical analysis and empirical evaluation, we conclude that the model selection procedure needs to be guided by careful considerations regarding the observed data, the factors used for correction, and the structure of the data-generating process.
Document type Conference contribution
Note With supplementary file
Language English
Published at https://doi.org/10.48550/arXiv.2106.03783
Published at https://papers.nips.cc/paper/2021/hash/93661c10ed346f9692f4d512319799b3-Abstract.html
Other links https://www.proceedings.com/63069.html
Downloads
2106.03783 (Accepted author manuscript)
Supplementary materials
Permalink to this page
Back