+ All Categories
Home > Documents > 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show...

11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show...

Date post: 03-Jul-2019
Category:
Upload: phungngoc
View: 220 times
Download: 0 times
Share this document with a friend
87
11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise
Transcript
Page 1: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise

Page 2: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Worst Additive Noise

• We will show that in terms of the capacity of the system, the zero-mean

Gaussian noise is the worst additive noise given that the noise vector has

a fixed correlation matrix.

• The diagonal elements of the correlation matrix specify the power of the

individual noise variables.

• The other elements in the matrix give a characterization of the correlation

between the noise variables.

Page 3: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Worst Additive Noise

• We will show that in terms of the capacity of the system, the zero-mean

Gaussian noise is the worst additive noise given that the noise vector has

a fixed correlation matrix.

• The diagonal elements of the correlation matrix specify the power of the

individual noise variables.

• The other elements in the matrix give a characterization of the correlation

between the noise variables.

Page 4: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Worst Additive Noise

• We will show that in terms of the capacity of the system, the zero-mean

Gaussian noise is the worst additive noise given that the noise vector has

a fixed correlation matrix.

• The diagonal elements of the correlation matrix specify the power of the

individual noise variables.

• The other elements in the matrix give a characterization of the correlation

between the noise variables.

Page 5: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Worst Additive Noise

• We will show that in terms of the capacity of the system, the zero-mean

Gaussian noise is the worst additive noise given that the noise vector has

a fixed correlation matrix.

• The diagonal elements of the correlation matrix specify the power of the

individual noise variables.

• The other elements in the matrix give a characterization of the correlation

between the noise variables.

Page 6: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 7: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 8: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 9: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 10: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 11: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 12: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 13: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 14: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 15: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 16: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 17: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 18: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 19: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 20: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 21: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 22: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 23: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

___________________

___________________

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 24: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

__________________________

__________________________

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 25: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

_________________

_________________

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 26: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Zero-Mean Gaussian System

1. Consider a system of correlated Gaussian channels

with noise vector Z

⇤ ⇠ N(0, K), and so

˜KZ

⇤ = K.

2. Let C⇤be the capacity of the system.

3. Let X

⇤be the zero-mean Gaussian input vector that

achieves the capacity.

4. Let Y

⇤be the output of the system with X

⇤as

input, i.e.,

Y

⇤= X

⇤+ Z

⇤.

Alternative System

1. Consider a system exactly the same as the Zero-

Mean Gaussian System except that the noise vector

Z, which has the same correlation matrix as Z

⇤, may

neither be zero-mean nor Gaussian.

2. Assume that the joint pdf of Z exists.

3. Let C be the capacity of the system.

4. Let Y be the output of the system with X

⇤as input,

i.e.,

Y = X

⇤+ Z.

Main Idea

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

• For the Zero-Mean Gaussian System, C⇤ = I(X⇤;Y⇤).

• For the alternative system, I(X⇤;Y) C.

• We will show that I(X⇤;Y⇤) I(X⇤;Y).

• Hence,C⇤ = I(X⇤;Y⇤) I(X⇤;Y) C.

Page 27: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if and only if Z = Z

⇤ ⇠ N (0, K).

Page 28: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Lemma 11.33 Let X be a zero-mean random vector and

Y = X+ Z

where Z is independent of X. Then

˜KY =

˜KX +

˜KZ.

Remark The scalar case has been proved in the proof of Theorem 11.21.

Lemma 11.34 LetY⇤ ⇠ N (0,K) andY be any random vector with correlation

matrix K. Then

ZfY⇤

(y) log fY⇤(y)dy =

Z

SY

fY(y) log fY⇤(y)dy.

Remark A similar technique has been used in proving Theorems 2.50 and

10.41 (maximum entropy distributions).

Page 29: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Lemma 11.33 Let X be a zero-mean random vector and

Y = X+ Z

where Z is independent of X. Then

˜KY =

˜KX +

˜KZ.

Remark The scalar case has been proved in the proof of Theorem 11.21.

Lemma 11.34 LetY⇤ ⇠ N (0,K) andY be any random vector with correlation

matrix K. Then

ZfY⇤

(y) log fY⇤(y)dy =

Z

SY

fY(y) log fY⇤(y)dy.

Remark A similar technique has been used in proving Theorems 2.50 and

10.41 (maximum entropy distributions).

Page 30: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Lemma 11.33 Let X be a zero-mean random vector and

Y = X+ Z

where Z is independent of X. Then

˜KY =

˜KX +

˜KZ.

Remark The scalar case has been proved in the proof of Theorem 11.21.

Lemma 11.34 LetY⇤ ⇠ N (0,K) andY be any random vector with correlation

matrix K. Then

ZfY⇤

(y) log fY⇤(y)dy =

Z

SY

fY(y) log fY⇤(y)dy.

Remark A similar technique has been used in proving Theorems 2.50 and

10.41 (maximum entropy distributions).

Page 31: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Lemma 11.33 Let X be a zero-mean random vector and

Y = X+ Z

where Z is independent of X. Then

˜KY =

˜KX +

˜KZ.

Remark The scalar case has been proved in the proof of Theorem 11.21.

Lemma 11.34 LetY⇤ ⇠ N (0,K) andY be any random vector with correlation

matrix K. Then

ZfY⇤

(y) log fY⇤(y)dy =

Z

SY

fY(y) log fY⇤(y)dy.

Remark A similar technique has been used in proving Theorems 2.50 and

10.41 (maximum entropy distributions).

Page 32: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 33: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 34: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 35: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 36: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 37: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 38: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 39: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 40: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 41: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 42: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 43: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 44: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 45: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 2.50 Let

p⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

X

x2Sp

p(x)ri(x) = ai for 1 i m. (1)

Then p⇤ maximizes H(p) over all probability distribu-

tion p on S subject to (1).

Sketch of Proof

H(p⇤) � H(p)

= �X

x2Sp⇤(x) ln p

⇤(x) +

X

x2Sp

p(x) ln p(x)

= �X

x2Sp

p(x) ln p⇤(x) +

X

x2Sp

p(x) ln p(x)

=

X

x2Sp

p(x) ln

p(x)

p⇤(x)

= D(pkp⇤)

� 0.

Remark The key step is to establish that

X

x2Sp⇤(x) ln p

⇤(x) =

X

x2Sp

p(x) ln p⇤(x).

Theorem 10.41 Let

f⇤(x) = e

��0

�Pm

i=1

�iri(x)

for all x 2 S, where �0

,�1

, · · · ,�m are chosen such

that

Z

Sf

ri(x)f(x)dx = ai for 1 i m. (2)

Then f⇤maximizes h(f) over all pdf f defined on S,

subject to the constraints in (2).

Remark The key step is to establish that

Z

Sf⇤(x) ln f

⇤(x)dx =

Z

Sf

f(x) ln f⇤(x)dx. (3)

Theorem 10.45 Let X be a vector of n continuous

random variables with correlation matrix

˜K. Then

h(X) 1

2

log

h(2⇡e)

n| ˜K|i

with equality if and only if X ⇠ N(0, ˜K).

Then (3) and Theorem 10.45 together imply

Lemma 11.34 Let Y

⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY

⇤ (y) log fY

⇤ (y)dy =

Z

SY

fY

(y) log fY

⇤ (y)dy.

Page 46: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Data Processing Inequality for Informational Divergence

Theorem Let X,X 0, Y , and Y 0be real random variables such that fXY and

fX0Y 0exist, and fY |X = fY 0|X0

. Then

D(fXkfX0) � D(fY kfY 0

).

Proof Exercise.

fY |XX

X 0 Y 0

Y

fY |X

Page 47: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Data Processing Inequality for Informational Divergence

Theorem Let X,X 0, Y , and Y 0be real random variables such that fXY and

fX0Y 0exist, and fY |X = fY 0|X0

. Then

D(fXkfX0) � D(fY kfY 0

).

Proof Exercise.

fY |XX

X 0 Y 0

Y

fY |X

Page 48: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Data Processing Inequality for Informational Divergence

Theorem Let X,X 0, Y , and Y 0be real random variables such that fXY and

fX0Y 0exist, and fY |X = fY 0|X0

. Then

D(fXkfX0) � D(fY kfY 0

).

Proof Exercise.

fY |XX

X 0 Y 0

Y

fY |X

Page 49: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Data Processing Inequality for Informational Divergence

Theorem Let X,X 0, Y , and Y 0be real random variables such that fXY and

fX0Y 0exist, and fY |X = fY 0|X0

. Then

D(fXkfX0) � D(fY kfY 0

).

Proof Exercise.

fY |XX

X 0 Y 0

Y

fY |X

Page 50: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Page 51: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Z� � N (0, K)

+X⇤ Y⇤

Page 52: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

Page 53: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

Page 54: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

Page 55: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.33 Let X be a zero-mean random vector

and

Y = X + Z

where Z is independent of X. Then

˜KY =

˜KX +

˜KZ.

Z� � N (0, K)

+X⇤ Y⇤

Z : K̃Z = K

+X⇤ Y

Page 56: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Page 57: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Page 58: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Page 59: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_____

Page 60: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_____

_____________________

Page 61: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_______

Page 62: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_______

___________________

Page 63: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

______

Page 64: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

______

__________________

Page 65: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

____

Page 66: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

____

____________________

Page 67: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.34 Let Y⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY⇤ (y) log fY⇤ (y)dy =

Z

SYfY(y) log fY⇤ (y)dy.

Page 68: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.34 Let Y⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY⇤ (y) log fY⇤ (y)dy =

Z

SYfY(y) log fY⇤ (y)dy.

______

Page 69: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.34 Let Y⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY⇤ (y) log fY⇤ (y)dy =

Z

SYfY(y) log fY⇤ (y)dy.

______

_____

Page 70: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.34 Let Y⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY⇤ (y) log fY⇤ (y)dy =

Z

SYfY(y) log fY⇤ (y)dy.

______

Page 71: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Lemma 11.34 Let Y⇤ ⇠ N(0, K) and Y be any ran-

dom vector with correlation matrix K. Then

ZfY⇤ (y) log fY⇤ (y)dy =

Z

SYfY(y) log fY⇤ (y)dy.

______

_____

Page 72: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

____________________

Page 73: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

____________________

__________________

Page 74: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

____________________

__________________

_____________________

Page 75: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

___________________

Page 76: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

___________________

____________________

Page 77: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

___________________

____________________

________________________

Page 78: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_____________________

Page 79: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_____________________

__________

Page 80: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_______________________

Page 81: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

_______________________

__________

Page 82: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

Page 83: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

+ Y

X⇤

Z

Page 84: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

+ Y

X⇤

Z

+ Y⇤

X⇤

Z⇤

Page 85: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

+ Y

X⇤

Z

+ Y⇤

X⇤

Z⇤

Page 86: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

+ Y

X⇤

Z

+ Y⇤

X⇤

Z⇤

Page 87: 11.9 Zero-Mean Gaussian Noise is the Worst Additive Noise · Worst Additive Noise • We will show that in terms of the capacity of the system, the zero-mean Gaussian noise is the

Theorem 11.32 For a fixed zero-mean

Gaussian random vector X

⇤, let

Y = X

⇤+ Z,

where the joint pdf of Z exists and Z is

independent of X

⇤. Under the constraint

that the correlation matrix of Z is equal

to K, where K is any symmetric positive

definite matrix, I(X

⇤;Y) is minimized if

and only if Z = Z

⇤ ⇠ N(0, K).

Proof

1. Since EZ

⇤= 0,

˜KZ

⇤ = KZ

⇤ = K. Therefore, Z

⇤and Z have the same correlation matrix.

2. By noting that X

⇤has zero mean, we apply

Lemma 11.33 to see that Y

⇤and Y have the same

correlation matrix.

3. The theorem is proved by considering

I(X

⇤;Y

⇤) � I(X

⇤;Y)

= h(Y

⇤) � h(Z

⇤) � h(Y) + h(Z)

= �Z

fY

⇤ (y) log fY

⇤ (y)dy +

ZfZ

⇤ (z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

= �Z

fY

(y) log fY

⇤ (y)dy +

Z

SZ

fZ

(z) log fZ

⇤ (z)dz

+

ZfY

(y) log fY

(y)dy �Z

SZ

fZ

(z) log fZ

(z)dz

=

Zlog

0

@fY

(y)

fY

⇤ (y)

1

A fY

(y)dy �Z

SZ

log

0

@fZ

(z)

fZ

⇤ (z)

1

A fZ

(z)dz

= D(fY

kfY

⇤ ) � D(fZ

kfZ

⇤ )

0.

+ Y

X⇤

Z

+ Y⇤

X⇤

Z⇤


Recommended