Differential entropy

The differential entropy of a continuous random variable X with probability density function (pdf) p(x) is

h(X)=p(x)logp(x)dx.

It is the continuous analogue of Shannon entropy for discrete distributions,

H(X)=ipilogpi.

Characteristics:

h(X)h(X)+log|a|. HΔ(X)h(X)+log1Δ.

Thus, differential entropy is not absolute: it only gains physical meaning when a fundamental resolution scale is specified.

Planck Length as a Natural Cutoff

In physics, the Planck length

P=Gc31.6×1035m

is often regarded as the minimal meaningful length scale. When interpreting entropy in continuous systems (fields, spacetime degrees of freedom, black holes), the bin size Δ can be taken to be on the order of P.
This gives a bridge between differential entropy and a physically grounded discrete Shannon entropy:

H(X)h(X)+log1P.

Gibbs entropy

Consider a system in Classical Statistical Mechanics with ensemble given by ρ.
The statistical definition of entropy, known as the Gibbs entropy, is

S=klnρ=kρ(q,p)ln[ρ(q,p)]dqdp

Important caution:

In some sense, S counts the number of microstates compatible with the macrostate. For a macrostate with ensemble a Dirac delta, there is only 1 compatible microstate, and the entropy diverges to .

Important relation

Consider a system in thermal equilibrium with a bath at temperature T described by the canonical ensemble, where the probability density is ρS(q,p)=1ZeβHS(q,p) with β=1/(kT).

  1. Express Entropy using the partition function (Z)
    Substituting ρS into the entropy definition:lnρS=ln(eβHSZ)=βHSlnZTaking the average:lnρS=βHSlnZPlugging into S=klnρS:S=kβE+klnZWith β=1/(kT):S=ET+klnZ
  2. Take the Differential
    For fixed Hamiltonian parameters:dSdβ=kβdEdβHence:dS=kβdESubstituting β=1/(kT):dE=TdS
Related: first law of thermodynamics.

Boltzmann entropy

See harmonic oscillator in CSM#5) Entropy


Maximum Entropy Distributions

p(x)=1μex/μ. p(x)=12πσ2exp((xμ)22σ2).

Derivation sketch (Lagrange multipliers)

Maximize Shannon entropy

H[p]=p(x)lnp(x)dx

subject to normalization and the moment constraints. The variational problem

L[p]=plnpdx+λ0(pdx1)+λ1(xpdxμ)+λ2((xμ)2pdxσ2)

yields (after variation)

p(x)exp(λ1xλ2(xμ)2).

In general, maximum entropy solutions belong to exponential families, with form determined by the active constraints.