Mutual Information

Sources:

  1. Elements of Information Theory
  2. An Introduction to Single-User Information Theory

Definition

Given two random variables X and Y, we want to define a measure of the information that Y provides about X when Y is observed, but X is not. We call this measure mutual information, which is defined as: I(X;Y)H(X)H(XY)

It can be expressed in terms relative entropy between their joint distribution pX,Y and the product of their marginal distributions pXpY I(X;Y)=x,yp(x,y)logpX,Y(x,y)pX(x)pY(y)=D(pX,YpXpY).

Properties of Mutual Information

Basics

  1. I(X;Y)=xXyYPX,Y(x,y)log2PX,Y(x,y)PX(x)PY(y).

  2. I(X;Y)=I(Y;X)=H(Y)H(YX).

  3. I(X;Y)=H(X)+H(Y)H(X,Y).

  4. I(X;Y)H(X) with equality holding iff X is a function of Y (i.e., X=f(Y) for some function f()).

  5. I(X;Y)0 with equality holding iff X and Y are independent.

  6. I(X;Y)min{log2|X|,log2|Y|}.

Proof:

Properties 1, 2, 3, and 4 follow immediately from the definition.

Property 5 is a direct consequence of D(pq)0, since I(X:Y)=D(pX,YpXpY) .

Property 6 holds iff I(X;Y)log2|X| and I(X;Y)log2|Y|.

To show the first inequality, we write I(X;Y)=H(X) H(XY), use the fact that H(XY) is nonnegative and apply Theorem: H(X)|X|. A similar proof can be used to show that I(X;Y)log2|Y|.

The relationships between H(X),H(Y),H(X,Y),H(XY),H(YX), and I(X;Y) can be illustrated by the Venn diagram in the above figure.

I(X:Y)=I(Y:X)

Expanding H(X)H(XY), we have: H(X)H(XY)=E[log1p(X)]E[log1p(XY)]=E[logp(XY)p(X)]=E[logp(XY)p(Y)p(X)p(Y)]=E[logp(X,Y)p(X)p(Y)]=H(Y)H(YX)

Then I(X;Y)H(X)H(XY)=H(Y)H(YX)

So mutual information is symmetric.

I(X;X)=H(X)

Now let’s ask an interesting question: How much does X tell us about itself? In other words, what is I (X; X)? Using our first definition, we have: I(X;X)=H(X)H(X|X)

We note that H(XX)=0, because in the expectation, X can only take on one fixed, given value with probability 1 . Therefore, H(XX)=log1=0. Thus: I(X;X)=H(X) Meaning that X tells us everything about itself!

I(X;Y)0

For any two random variables, X,Y, I(X;Y)0 with equality if and only if X and Y are independent.

Proof:

  1. We know that I(X;Y)=D(p(x,y)p(x)p(y)).
  2. See property 3 of Relative Entropy, D(pq)0 for all distributions p,q with equality holding iff p=q.
  3. So I(X;Y)=D(p(x,y)p(x)p(y))0, with equality if and only if p(x,y)=p(x)p(y) (i.e., X and Y are independent).

Corollary: I(X;YZ)0, with equality if and only if X and Y are conditionally independent given Z.

Conditional Mutual Information

img

Venn diagram of information theoretic measures for three variables x,y, and z, represented by the lower left, lower right, and upper circles, respectively.

The conditional mutual informations I(x;zy),I(y;zx) and I(x;yz) are represented by the yellow, cyan, and magenta regions, respectively.

The conditional mutual information, denoted by I(X;YZ), is defined as the common uncertainty between X and Y under the knowledge of Z : I(X;YZ):=H(XZ)H(XY,Z)

可以这么想像: I(X;Y) 就是 H(X)H(Y) 的交集, 再挖掉其中 H(Z) 的部分就是I(X;Y|Z). 对应于图中粉色部分.

Joint Mutual Information

Lemma: Defining the joint mutual information between X and the pair (Y,Z) by I(X;Y,Z):=H(X)H(XY,Z), we have I(X;Y,Z)=I(X;Y)+I(X;ZY)=I(X;Z)+I(X;YZ). X(Y,Z)的互信息 = XY 的互信息 + 在Y 已知的情况下 XZ 的互信息.

可以这么想像: 把H(Y),H(Z) 连成一块得到H(Y,Z), I(X;Y,Z) 就是 H(X)H(Y,Z) 的交集. 对应于图中黄, 灰, 粉三块区域的并集.

Proof: Without loss of generality, we only prove the first equality: I(X;Y,Z)=H(X)H(XY,Z)=H(X)H(XY)+H(XY)H(XY,Z)=I(X;Y)+I(X;ZY).

The above lemma can be read as follows: the information that (Y,Z) has about X is equal to the information that Y has about X plus the information that Z has about X when Y is already known.

Properties of Entropy and Mutual Information for Multiple Random Variables

Chain rule for joint entropy

Theorem: Let X1,X2,,Xn be drawn according to PXn(xn):=PX1,,Xn(x1,,xn), where we use the common superscript notation to denote an n-tuple: Xn:=(X1,,Xn) and xn:=(x1,,xn).

Then H(X1,X2,,Xn)=i=1nH(XiXi1,,X1), where H(XiXi1,,X1):=H(X1) for i=1. (The above chain rule can also be written as: H(Xn)=i=1nH(XiXi1) where Xi:=(X1,,Xi).)

For example, for three random variables X, Y , and Z, H(X,Y,Z)=H(X)+H(Y,ZX)=H(X)+H(YX)+H(ZX,Y)


Proof:

From chain rule for 2 r.v. , H(X1,X2,,Xn)=H(X1,X2,,Xn1)+H(XnXn1,,X1)

Once again, applying chain rule for 2 r.v. to the first term of the right-hand side of this equation, we have H(X1,X2,,Xn1)=H(X1,X2,,Xn2)+H(Xn1Xn2,,X1)

The desired result can then be obtained by repeatedly applying chain rule for 2 r.v. .

Chain rule for conditional entropy

Theorem: H(X1,X2,,XnY)=i=1nH(XiXi1,,X1,Y).


Proof:

The theorem can be proved similarly to Chain Rule for Entropy(2 Variables). If Xn=(X1,,Xn) and Ym=(Y1,,Ym) are jointly distributed random vectors (of not necessarily equal lengths), then their joint mutual information is given by I(X1,,Xn;Y1,,Ym):=H(X1,,Xn)H(X1,,XnY1,,Ym).

Chain rule for mutual information

Theorem: I(X1,X2,,Xn;Y)=i=1nI(Xi;YXi1,,X1), where I(Xi;YXi1,,X1):=I(X1;Y) for i=1.


Proof:

This can be proved by first expressing mutual information in terms of entropy and conditional entropy, and then applying the chain rules for entropy and conditional entropy.

Independence bound on entropy

Theorem: H(X1,X2,,Xn)i=1nH(Xi).

Equality holds iff all the Xi 's are independent of each other.[^8]


Proof:

By applying the chain rule for entropy, H(X1,X2,,Xn)=i=1nH(XiXi1,,X1)i=1nH(Xi).

Equality holds iff each conditional entropy is equal to its associated entropy, that iff Xi is independent of (Xi1,,X1) for all i.

Bound on mutual information

Theorem: If {(Xi,Yi)}i=1n is a process satisfying the conditional independence assumption PYnXn=i=1nPYiXi, then I(X1,,Xn;Y1,,Yn)i=1nI(Xi;Yi), with equality holding iff {Xi}i=1n are independent.


Proof:

From the independence bound on entropy, we have H(Y1,,Yn)i=1nH(Yi).

By the conditional independence assumption, we have H(Y1,,YnX1,,Xn)=E[log2PYnXn(YnXn)]=E[i=1nlog2PYiXi(YiXi)]=i=1nH(YiXi).

Hence, I(Xn;Yn)=H(Yn)H(YnXn)i=1nH(Yi)i=1nH(YiXi)=i=1nI(Xi;Yi), with equality holding iff {Yi}i=1n are independent, which holds iff {Xi}i=1n are independent.