Cross Mutual Information

📅 2025-07-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Quantifying and comparing statistical dependence between random variables X and Y across distinct sample sets is challenging under non-stationary distributions, as conventional dependence measures—such as mutual information—lack comparability under distributional shift. Method: We propose Cross Mutual Information (CMI), the first formally defined, cross-sample comparable dependence measure. CMI jointly estimates the joint and marginal distributions from two independent samples to assess consistency in the X–Y dependence structure, thereby overcoming the fundamental incomparability of standard mutual information under distribution shift. Contribution/Results: We establish a tight theoretical connection between CMI and the coefficient of determination (R²) in linear regression. In extensive simulations across diverse nonlinear dependence structures, CMI consistently outperforms baseline methods and exhibits robustness to distribution drift. This framework provides an interpretable, generalizable tool for cross-subject or cross-task dependence modeling in high-dimensional non-stationary domains such as neuroimaging.

Technology Category

Application Category

📝 Abstract
Mutual information (MI) is a useful information-theoretic measure to quantify the statistical dependence between two random variables: $X$ and $Y$. Often, we are interested in understanding how the dependence between $X$ and $Y$ in one set of samples compares to another. Although the dependence between $X$ and $Y$ in each set of samples can be measured separately using MI, these estimates cannot be compared directly if they are based on samples from a non-stationary distribution. Here, we propose an alternative measure for characterising how the dependence between $X$ and $Y$ as defined by one set of samples is expressed in another, extit{cross mutual information}. We present a comprehensive set of simulation studies sampling data with $X$-$Y$ dependencies to explore this measure. Finally, we discuss how this relates to measures of model fit in linear regression, and some future applications in neuroimaging data analysis.
Problem

Research questions and friction points this paper is trying to address.

Quantify dependence between X and Y in different sample sets
Compare MI estimates from non-stationary distributions effectively
Propose cross mutual information for cross-sample dependency analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes cross mutual information measure
Compares dependencies in non-stationary distributions
Applies to neuroimaging data analysis
🔎 Similar Papers
No similar papers found.
Chetan Gohil
Chetan Gohil
University of Sydney
Computational NeuroscienceAccelerator Physics
O
Oliver M Cliff
Centre for Complex Systems, The University of Sydney
James M Shine
James M Shine
Neuroimaging Research Fellow, The University of Sydney
Neuroscience
B
Ben D Fulcher
School of Physics, The University of Sydney
J
Joseph T Lizier
Centre for Complex Systems, The University of Sydney