Nonlinear Measures of Statistical Dependence (Maximum Information Dimension)
Measuring statistical dependence between two random variables
Summary
An estimation algorithm for MID (Mutual Information Dimension), which measures statistical dependence between two random variables and produces a real-valued score from 0 (weak) to 1 (strong). This algorithm has the following advantages:
- Nonlinear dependencies (and also linear dependences) can be measured,
- Scalable; the average-case time complexity is O(nlogn), where n is the number of data points, and
- Parameter-free.
Code
C implementation can be downloaded here: Download code.zip (ZIP, 421 KB)
Code is also available at external page GitHub
Further information and publication
Please see the following article for detailed information and refer it in your published research.
Measuring Statistical Dependence via the Mutual Information Dimension
Mahito Sugiyama and Karsten Borgwardt
Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI 2013), 1692-1698
external page Online | ETH Research Collection | Project page | external page GitHub
Contact: external page Mahito Sugiyama