Table of Contents
- What Is Information?
- Entropy of Discrete Variables
- The Source Coding Theorem
- The Noisy Channel Coding Theorem
- Entropy of Continuous Variables
- Mutual Information: Continuous
- Channel Capacity: Continuous
- Rate Distortion Theory
- Transfer Entropy
- Thermodynamic Entropy and Information
- Information As Nature's Currency

