An exploration of the role of principal inertia components in information theory

The principal inertia components of the joint distribution of two random variables X and Y are inherently connected to how an observation of Y is statistically related to a hidden variable X. In this paper, we explore this connection within an information theoretic framework. We show that, under cer...

Full description

Bibliographic Details
Main Authors: Calmon, Flavio P., Varia, Mayank, Medard, Muriel
Other Authors: Lincoln Laboratory
Format: Article
Language:en_US
Published: Institute of Electrical and Electronics Engineers (IEEE) 2016
Online Access:http://hdl.handle.net/1721.1/100951
https://orcid.org/0000-0003-2912-7972
https://orcid.org/0000-0003-4059-407X
Description
Summary:The principal inertia components of the joint distribution of two random variables X and Y are inherently connected to how an observation of Y is statistically related to a hidden variable X. In this paper, we explore this connection within an information theoretic framework. We show that, under certain symmetry conditions, the principal inertia components play an important role in estimating one-bit functions of X, namely f(X), given an observation of Y. In particular, the principal inertia components bear an interpretation as filter coefficients in the linear transformation of p[subscript f(X)|X] into p[subscript f(X)|Y]. This interpretation naturally leads to the conjecture that the mutual information between f(X) and Y is maximized when all the principal inertia components have equal value. We also study the role of the principal inertia components in the Markov chain B → X → Y → B̂, where B and B̂ are binary random variables. We illustrate our results for the setting where X and Y are binary strings and Y is the result of sending X through an additive noise binary channel.