Figure 1
Placeholder for a particular trick in information theory:
Williams and Beer (2010) introduced partial information decomposition (PID) as a way to split the mutual information that a set of sources has about a target into nonânegative âatomsâ corresponding to redundant, unique, and synergistic information. Their framework has become a standard reference point, but their specific redundancy measure has been heavily critiqued. There are now many alternative proposals and generalizations.
Williams & Beerâs original PID
Williams & Beer consider a target (Y) and sources (X_1, X_2, \dots), and aim to decompose (I(X_1, X_2, \dots : Y)) into atoms corresponding to:
**âŚ
Figure 1
Placeholder for a particular trick in information theory:
Williams and Beer (2010) introduced partial information decomposition (PID) as a way to split the mutual information that a set of sources has about a target into nonânegative âatomsâ corresponding to redundant, unique, and synergistic information. Their framework has become a standard reference point, but their specific redundancy measure has been heavily critiqued. There are now many alternative proposals and generalizations.
Williams & Beerâs original PID
Williams & Beer consider a target (Y) and sources (X_1, X_2, \dots), and aim to decompose (I(X_1, X_2, \dots : Y)) into atoms corresponding to:
Redundant information (shared by multiple sources about (Y))
Unique information (available only from one source)
Synergistic information (available only from sources in combination).
They formalize this by:
Introducing a redundancy function (I_{\cap}(X_{1:n} : Y)) defined on sets of sources.
Requiring this redundancy measure to satisfy three axioms: symmetry, selfâredundancy, and monotonicity.
Using the lattice of âinformation antichainsâ over source subsets, they show that, given such a redundancy measure, the mutual information can be decomposed by Moebius inversion over this lattice, yielding guaranteed nonânegative information atoms.
As a concrete proposal, they define redundancy via the âminimum informationâ measure (I_{\min}), which roughly takes the minimal specific information any source provides about each outcome of (Y), averaged over outcomes.
Subsequent theoretical developments
While the framework (axioms + lattice) was widely accepted, the specific WilliamsâBeer redundancy (I_{\min}) quickly attracted criticism.
Much subsequent work has tried to preserve the WilliamsâBeer axioms and lattice structure while replacing (I_{\min}) with betterâbehaved redundancy or intersectionâinformation measures.
References
Friedman, Mosenzon, Slonim, et al. 2001. âMultivariate Information Bottleneck.â In Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence. UAIâ01.
MartĂnez-SĂĄnchez, Arranz, and Lozano-DurĂĄn. 2024. âDecomposing Causality into Its Synergistic, Unique, and Redundant Components.â Nature Communications.
StudenĂ˝, and VejnarovĂĄ. 1998. âOn Multiinformation Function as a Tool for Measuring Stochastic Dependence.â In Learning in Graphical Models.