Bias in O-Information Estimation

Open Access
Authors
Publication date 10-2024
Journal Entropy
Article number 837
Volume | Issue number 26 | 10
Number of pages 24
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Faculty of Social and Behavioural Sciences (FMG) - Psychology Research Institute (PsyRes)
Abstract
Higher-order relationships are a central concept in the science of complex systems. A popular method of attempting to estimate the higher-order relationships of synergy and redundancy from data is through the O-information. It is an information–theoretic measure composed of Shannon entropy terms that quantifies the balance between redundancy and synergy in a system. However, bias is not yet taken into account in the estimation of the O-information of discrete variables. In this paper, we explain where this bias comes from and explore it for fully synergistic, fully redundant, and fully independent simulated systems of (Formula presented.) variables. Specifically, we explore how the sample size and number of bins affect the bias in the O-information estimation. The main finding is that the O-information of independent systems is severely biased towards synergy if the sample size is smaller than the number of jointly possible observations. This could mean that triplets identified as highly synergistic may in fact be close to independent. A bias approximation based on the Miller–Maddow method is derived for the O-information. We find that for systems of (Formula presented.) variables the bias approximation can partially correct for the bias. However, simulations of fully independent systems are still required as null models to provide a benchmark of the bias of the O-information.
Document type Article
Language English
Published at https://doi.org/10.3390/e26100837
Other links https://www.scopus.com/pages/publications/85207669127
Downloads
Bias in O-Information Estimation (Final published version)
Permalink to this page
Back