Given a set of N variables and a set of M samples representing information on each of the variables, how do we find functional relationships between the variables? MIST tackles the problem by using a divide-and-conquer approach: while considering the set of all variable tuples as the search space, MIST divides that search space among parallel threads, and then conquers it by computing dependency measures for each tuple.
Symmetric Delta (Galas et al, 2020; Galas et al, 2014 (link)) is the measure used in a MIST search. The Symmetric Delta is a novel symmetric measure of functional dependence (it is symmetric under exchange of variables) constructed from joint entropies. Joint entropies between variables (using the Shannon entropy (Shannon and Weaver, 1949 ) defined as the expectation of the logarithm of each element of the joint probability distribution: HX=iPxilogPxi ). For variable tuples of size T there are N choose T tuples in the search space. Thus, the problem is reduced to computing joint probability distributions for a very large number of variable tuples.
Dependence between two variables X and Y can be directly measured with mutual information IX,Y , defined as
IX,Y=HX+HYHX,Y,
where HX and HY are single entropies of variables X and Y and HX,Y is their joint entropy.
A general dependence among three variables, X, Y, and Z, can be measured with symmetric delta. Δ¯X,Y,Z . To see clearly the definition of symmetric delta, we need to introduce interaction information, which is a multivariable generalization of mutual information (McGill, 1954 ), defined for three variables as
IX,Y,Z=IX,YIX,Y|Z.
Given interaction information, differential interaction information Δ is defined as a difference between values of successive interaction information arising from adding a variable:
ΔX=IX,Y,ZIY,Z,ΔY=IX,Y,ZIX,Z,ΔZ=IX,Y,ZIX,Y.
Here ΔX is called asymmetric delta for the target variable X. To detect a fully synergistic dependence among a set of variables, we want a single measure, which is symmetric. Consequently, we defined a general measure Δ¯ , called symmetric delta (or simply delta), by multiplying the Δ 's with all possible choices of the target variable:
Δ¯X,Y,Z=ΔXΔYΔZ.
The critical property of this delta measure is that it is zero whenever any of the three variables is independent of the others. It is important to note that the absolute values of the delta measure indicate the degree to which the corresponding variables are collectively interdependent.
Free full text: Click here