Classical Statistical Mechanics

Key example: Harmonic oscillator in CSM.
In the 19th century, a central challenge in theoretical Physics was describing systems with an enormous number of degrees of freedom (e.g., 1023 particles), as required to derive thermodynamics from microscopic laws. Solving such systems deterministically—via a Cauchy problem with 1023 initial conditions—was hopeless. Instead, a radical shift occurred: rather than tracking individual phase-space trajectories, physicists began describing states probabilistically, using probability distributions over phase space. This introduced randomness into fundamental physics, not as a limitation of knowledge, but as a necessary feature of statistical mechanics.

Scope and limitations:

System

In general, you consider a system as a collection of states X. We need X to be a measure space (X,μ). Typically is a phase space of a classical Hamiltonian system, with the Liouville measure.

Energy

You have an energy function E:XR. Typically is the Hamiltonian of a classical Hamiltonian system.

Volume

You can also have another function V:XR, called volume. See first law of thermodynamics.

Microstates and macrostates

The points xX are called microstates. But we usually do not have access to measurements of particular microstates, only to macroscopic variables (typically mean temperature, mean energy, total energy, …). These observables impose constraints on the microstates.

The collection of macroscopic constraints, together with any assumptions used to determine the probability distribution (such as equal a priori probability or maximum entropy), is called a macrostate.

Ensembles

This brings us to the notion of an ensemble: a probability distribution over microstates consistent with the macrostate information.

Entropy

There are many ensembles compatible with a given macrostate. To choose a single, unbiased ensemble from the many possibilities compatible with a macrostate, we invoke the Principle of Maximum Entropy. This principle states that the best choice for ρ(x) is the one that maximizes the Gibbs entropy (defined below) subject to the known macroscopic constraints. This ensures that we don't assume any information we don't have.
The ensembles have associated a Gibbs/Shannon entropy:

S(ρ)=ρln(ρ)μ,

or

S(ρ)=ρln(ρ)d3Nrd3Np,

in the case of a phase space of a classical Hamiltonian system.
It is interpreted as a measure of our ignorance of the exact microstate of the system. The higher the entropy, the less information we have about where exactly the system is in phase space. A sharply peaked ρ corresponds to low entropy (high knowledge of the system), while a broad ρ corresponds to high entropy (greater uncertainty).

This functional has the same form as Shannon entropy in information theory and is uniquely characterized by its additivity, continuity, and the fact that it is maximized by uniform distributions under constraints.

Following the principle of maximum entropy we obtain, for example:

ρ(x)={0, if E(x)3,1Ω, if E(x)=3,

where Ω is the measure of the set {xX:E(x)=3}. I think this kind of ensembles are called microcanonical ensembles. In the microcanonical ensemble, it reduces (up to constants) to Boltzmann’s formula S=klnΩ.

Liouville's equation

The Hamiltonian equations for a particle

pt=Hq,qt=Hp,

correspond to the ideal case of a particle perfectly localized in phase space and we can represent this situation by the "degenerated probability density" ρ(p,q,t)=δ(pp0(t))δ(qq0(t)). Now, if we build a general density ρ as superpositions of "δs" as ρ=Σipiδpi(t),qi(t), we may show, by Liouville's theorem, that any probability density ρ carried along the flow must satisfy the continuity equation

tρ+(ρXH)=0,

and since (ρXH)={ρ,H}={H,ρ}, one immediately obtains the Liouville equation

(6)tρ={H,ρ}.

Now, starting from a density ρ that verifies (6), we can look at the evolution of the expectation value ft of an observable f(p,q,t) defined as:

f(p,q,t)t=d3pd3qρ(q,p,t)f(p,q,t)

In turn, integrating ddtft=(tρ)f+ρtf by parts (with no boundary terms thanks to incompressibility) yields

ddtft=tft+{H,f}.

Related: