Abstract
This paper considers a general linear vector Gaussian channel with arbitrary signaling and pursues two closely related goals: i) closed-form expressions for the gradient of the mutual information with respect to arbitrary parameters of the system, and ii) fundamental connections between information theory and estimation theory. Generalizing the fundamental relationship recently unveiled by Guo, Shamai, and Verdú, we show that the gradient of the mutual information with respect to the channel matrix is equal to the product of the channel matrix and the error covariance matrix of the best estimate of the input given the output. Gradients and derivatives with respect to other parameters are then found via the differentiation chain rule.
Original language | English (US) |
---|---|
Pages (from-to) | 141-154 |
Number of pages | 14 |
Journal | IEEE Transactions on Information Theory |
Volume | 52 |
Issue number | 1 |
DOIs | |
State | Published - Jan 2006 |
All Science Journal Classification (ASJC) codes
- Information Systems
- Computer Science Applications
- Library and Information Sciences
Keywords
- De Bruijn's identity
- Divergence
- Gaussian noise
- Minimum mean-square error (MMSE)
- Multiple-input multiple-output (MIMO) channels
- Mutual information
- Nonlinear estimation
- Precoder optimization