In information theory, dual total correlation,[1] information rate,[2] excess entropy,[3][4] or binding information[5] is one of several known non-negative generalizations of mutual information. While total correlation is bounded by the sum entropies of the n elements, the dual total correlation is bounded by the joint-entropy of the n elements. Although well behaved, dual total correlation has received much less attention than the total correlation. A measure known as "TSE-complexity" defines a continuum between the total correlation and dual total correlation.[3]
For a set of n random variables { X 1 , … , X n } {\displaystyle \{X_{1},\ldots ,X_{n}\}} , the dual total correlation D ( X 1 , … , X n ) {\displaystyle D(X_{1},\ldots ,X_{n})} is given by
where H ( X 1 , … , X n ) {\displaystyle H(X_{1},\ldots ,X_{n})} is the joint entropy of the variable set { X 1 , … , X n } {\displaystyle \{X_{1},\ldots ,X_{n}\}} and H ( X i ∣ ⋯ ) {\displaystyle H(X_{i}\mid \cdots )} is the conditional entropy of variable X i {\displaystyle X_{i}} , given the rest.
The dual total correlation normalized between [0,1] is simply the dual total correlation divided by its maximum value H ( X 1 , … , X n ) {\displaystyle H(X_{1},\ldots ,X_{n})} ,
Dual total correlation is non-negative and bounded above by the joint entropy H ( X 1 , … , X n ) {\displaystyle H(X_{1},\ldots ,X_{n})} .
Secondly, Dual total correlation has a close relationship with total correlation, C ( X 1 , … , X n ) {\displaystyle C(X_{1},\ldots ,X_{n})} , and can be written in terms of differences between the total correlation of the whole, and all subsets of size N − 1 {\displaystyle N-1} :[6]
where X = { X 1 , … , X n } {\displaystyle {\textbf {X}}=\{X_{1},\ldots ,X_{n}\}} and X − i = { X 1 , … , X i − 1 , X i + 1 , … , X n } {\displaystyle {\textbf {X}}^{-i}=\{X_{1},\ldots ,X_{i-1},X_{i+1},\ldots ,X_{n}\}}
Furthermore, the total correlation and dual total correlation are related by the following bounds:
Finally, the difference between the total correlation and the dual total correlation defines a novel measure of higher-order information-sharing: the O-information:[7]
The O-information (first introduced as the "enigmatic information" by James and Crutchfield[8] is a signed measure that quantifies the extent to which the information in a multivariate random variable is dominated by synergistic interactions (in which case Ω ( X ) < 0 {\displaystyle \Omega ({\textbf {X}})<0} ) or redundant interactions (in which case Ω ( X ) > 0 {\displaystyle \Omega ({\textbf {X}})>0} .
Han (1978) originally defined the dual total correlation as,
However Abdallah and Plumbley (2010) showed its equivalence to the easier-to-understand form of the joint entropy minus the sum of conditional entropies via the following: