Entropy, Shannon’s Measure of Information and Boltzmann’s H-Theorem
We start with a clear distinction between Shannon’s Measure of Information (SMI) and the Thermodynamic Entropy. The first is defined on any probability distribution; and therefore it is a very general concept. On the other hand Entropy is defined on a very special set of distributions. Next we show...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2017-01-01
|
Series: | Entropy |
Subjects: | |
Online Access: | http://www.mdpi.com/1099-4300/19/2/48 |