Two Measures of Dependence.

Entropy (Basel)

Signal and Information Processing Laboratory, ETH Zurich, 8092 Zurich, Switzerland.

Published: August 2019

Two families of dependence measures between random variables are introduced. They are based on the Rényi divergence of order α and the relative α -entropy, respectively, and both dependence measures reduce to Shannon's mutual information when their order α is one. The first measure shares many properties with the mutual information, including the data-processing inequality, and can be related to the optimal error exponents in composite hypothesis testing. The second measure does not satisfy the data-processing inequality, but appears naturally in the context of distributed task encoding.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7515307PMC
http://dx.doi.org/10.3390/e21080778DOI Listing

Publication Analysis

Top Keywords

dependence measures
8
data-processing inequality
8
measures dependence
4
dependence families
4
families dependence
4
measures random
4
random variables
4
variables introduced
4
introduced based
4
based rényi
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!