User:XORnado/sandbox
Partial Information Decomposition is an extension of information theory, that aims to generalize the pairwise relations described by information theory to the interaction of multiple variables.[1]
Motivation
[edit]Information theory can quantify the amount of information a single source variable has about a target variable via the mutual information . If we now consider a second source variable , classical information theory can only describe the mutual information of the joint variable with , given by . In general however, it would be interesting to know how exactly the individual variables and and their interactions relate to .
Consider that we are given two source variables and a target variable . In this case the total mutual information , while the individual mutual information . That is, there is synergistic information arising from the interaction of about , which cannot be easily captured with classical information theoretic quantities.
Definition for two source variables
[edit]Partial information decomposition further decomposes the mutual information between the source variables with the target variable as
Here the individual components, called "information atoms", are defined as
- is the unique information that has about , which is not in
- is the unique information that has about , which is not in
- is the synergistic information that is in the interaction of and about
- is the redundant information that is in both or about
There is, thus far, no universal agreement on how these terms should be defined, with different approaches that decompose information into redundant, unique, and synergistic components appearing in the literature.[1][2][3][4]
General definition
[edit]A key principle for generalizing PID to source variables is the notion of containment of information within different subsets of those variables.[5] For , there are three relevant subsets of source indices: , , and . Every possible way of how information about the target could distributed over subsets of sources can be represented in a containment table, as shown below. This table also includes the empty set for completeness.
Type | ||||
---|---|---|---|---|
Redundancy | 0 | 1 | 1 | 1 |
Unique to | 0 | 1 | 0 | 1 |
Unique to | 0 | 0 | 1 | 1 |
Synergy | 0 | 0 | 0 | 1 |
In this table, a 1 indicates that the information atom is contained in that subset of source variables, while a 0 indicates that it is not. Three main constraints ensure the table captures meaningful information relationships:
- Empty-set constraint: There is no information in the empty set, so the first column must always be 0.
- Full-set constraint: All the information available from the sources is contained in the full set, so the last column must always be 1.
- Monotonicity: If an atom is contained in some subset of sources, it must also be contained in any superset of that subset (the table entries cannot “go back” from 1 to 0 as you move to larger sets).
To handle arbitrary sources, one constructs a larger containment table under these same constraints. Each row corresponds to an information atom characterized by how it is contained in the different subsets of source variables. Mathematically, each row can be seen as a monotonic Boolean function from the power set of sources to {0,1}, and the number of such functions is given by the Dedekind numbers (minus two, since the constant-zero and constant-one functions are prohibited by the first two constraints).
It is conventional to label an information atom by the smallest subsets (in terms of subset relations) from which it can be obtained. This is sufficient because due to the monotonicity constraint it is implied that the atom must be contained in all larger sets as well. For , for example:
- The redundancy atom is denoted .
- The unique atoms are and .
- The synergy atom is .
Generally, the sets listed in the argument of form a so called antichain, i.e. a collection of sets where no set is a subset of another one. A generic antichain is denoted by the letter and the information atom corresponding to it by .
The full containment table for showing 18 distinct atoms and their antichain representation is shown in the following collapsible table:
# | Atom | ||||||||
---|---|---|---|---|---|---|---|---|---|
1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | |
2 | 0 | 1 | 0 | 0 | 1 | 1 | 0 | 1 | |
3 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 1 | |
4 | 0 | 0 | 0 | 1 | 0 | 1 | 1 | 1 | |
5 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | |
6 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | |
7 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | |
8 | 0 | 1 | 1 | 0 | 1 | 1 | 1 | 1 | |
9 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | |
10 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | |
11 | 0 | 1 | 0 | 0 | 1 | 1 | 1 | 1 | |
12 | 0 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | |
13 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | |
14 | 0 | 0 | 0 | 0 | 1 | 1 | 0 | 1 | |
15 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 1 | |
16 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | |
17 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | |
18 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 |
For an arbitrary set of sources , a partial information decomposition is a collection of atoms , one per antichain, so that for every subset , the following condition is satisfied
where is the mutual information provided by the sources indexed by . Intuitively, this condition simply says that the mutual information provided by should be the sum of atoms that are contained in it, i.e. those with a 1 on in the containment table. By construction, these are exactly the atoms with a subset in their antichain representation.
For , the condition states that all atoms have to sum up to the joint mutual information. This is because the sets contained in any are always subsets of the full set and therefore the condition under the summation sign is always fulfilled.
The condition above specifies a linear system of equations relating the information atoms to mutual information. However, since the number of subsets of sources grows exponentially in while the number of information atoms grows super-exponentially, the system becomes more and more underdetermined as grows. Already for it provides three equations for four unknowns. Hence, there is no unique solution for the information atoms .
Various proposals have been made to resolve this underdetermination by imposing additional constraints on the information atoms. Typically, these constraints tie the atoms to other information-theoretic quantities besides the mutual information. For instance, the original approach by Williams and Beer [1] introduced a "redundancy function," while other authors have used "union information" or synergy functions. Importantly, these quantities are in general (i.e. for ) themselves composed of multiple information atoms and should not be confused with a single "redundancy atom" or "synergy atom."
Applications
[edit]Despite the lack of universal agreement, partial information decomposition has been applied to diverse fields, including climatology,[6] neuroscience[7][8][9] sociology,[10] and machine learning[11] Partial information decomposition has also been proposed as a possible foundation on which to build a mathematically robust definition of emergence in complex systems[12] and may be relevant to formal theories of consciousness.[13]
See also
[edit]References
[edit]- ^ a b c Williams PL, Beer RD (2010-04-14). "Nonnegative Decomposition of Multivariate Information". arXiv:1004.2515 [cs.IT].
- ^ Quax R, Har-Shemesh O, Sloot PM (February 2017). "Quantifying Synergistic Information Using Intermediate Stochastic Variables". Entropy. 19 (2): 85. arXiv:1602.01265. doi:10.3390/e19020085. ISSN 1099-4300.
- ^ Rosas FE, Mediano PA, Rassouli B, Barrett AB (2020-12-04). "An operational information decomposition via synergistic disclosure". Journal of Physics A: Mathematical and Theoretical. 53 (48): 485001. arXiv:2001.10387. Bibcode:2020JPhA...53V5001R. doi:10.1088/1751-8121/abb723. ISSN 1751-8113. S2CID 210932609.
- ^ Kolchinsky A (March 2022). "A Novel Approach to the Partial Information Decomposition". Entropy. 24 (3): 403. arXiv:1908.08642. Bibcode:2022Entrp..24..403K. doi:10.3390/e24030403. PMC 8947370. PMID 35327914.
- ^ Gutknecht AJ, Wibral M, Makkeh A (2021). "Bits and Pieces: Understanding Information Decomposition from Part-Whole Relationships and Formal Logic". Proceedings of the Royal Society A. 477 (2251): 20210110. arXiv:2101.03968. doi:10.1098/rspa.2021.0110.
- ^ Goodwell AE, Jiang P, Ruddell BL, Kumar P (February 2020). "Debates—Does Information Theory Provide a New Paradigm for Earth Science? Causality, Interaction, and Feedback". Water Resources Research. 56 (2). Bibcode:2020WRR....5624940G. doi:10.1029/2019WR024940. ISSN 0043-1397. S2CID 216201598.
- ^ Newman EL, Varley TF, Parakkattu VK, Sherrill SP, Beggs JM (July 2022). "Revealing the Dynamics of Neural Information Processing with Multivariate Information Decomposition". Entropy. 24 (7): 930. Bibcode:2022Entrp..24..930N. doi:10.3390/e24070930. PMC 9319160. PMID 35885153.
- ^ Luppi AI, Mediano PA, Rosas FE, Holland N, Fryer TD, O'Brien JT, et al. (June 2022). "A synergistic core for human brain evolution and cognition". Nature Neuroscience. 25 (6): 771–782. doi:10.1038/s41593-022-01070-0. PMC 7614771. PMID 35618951. S2CID 249096746.
- ^ Wibral M, Priesemann V, Kay JW, Lizier JT, Phillips WA (March 2017). "Partial information decomposition as a unified approach to the specification of neural goal functions". Brain and Cognition. Perspectives on Human Probabilistic Inferences and the 'Bayesian Brain'. 112: 25–38. arXiv:1510.00831. doi:10.1016/j.bandc.2015.09.004. PMID 26475739. S2CID 4394452.
- ^ Varley TF, Kaminski P (October 2022). "Untangling Synergistic Effects of Intersecting Social Identities with Partial Information Decomposition". Entropy. 24 (10): 1387. Bibcode:2022Entrp..24.1387V. doi:10.3390/e24101387. ISSN 1099-4300. PMC 9611752. PMID 37420406.
- ^ Tax TM, Mediano PA, Shanahan M (September 2017). "The Partial Information Decomposition of Generative Neural Network Models". Entropy. 19 (9): 474. Bibcode:2017Entrp..19..474T. doi:10.3390/e19090474. hdl:10044/1/50586. ISSN 1099-4300.
- ^ Mediano PA, Rosas FE, Luppi AI, Jensen HJ, Seth AK, Barrett AB, et al. (July 2022). "Greater than the parts: a review of the information decomposition approach to causal emergence". Philosophical Transactions. Series A, Mathematical, Physical, and Engineering Sciences. 380 (2227): 20210246. doi:10.1098/rsta.2021.0246. PMC 9125226. PMID 35599558.
- ^ Luppi AI, Mediano PA, Rosas FE, Harrison DJ, Carhart-Harris RL, Bor D, Stamatakis EA (2021). "What it is like to be a bit: an integrated information decomposition account of emergent mental phenomena". Neuroscience of Consciousness. 2021 (2): niab027. doi:10.1093/nc/niab027. PMC 8600547. PMID 34804593.