Classical Statistical Mechanics
Key example: Harmonic oscillator in CSM.
In the 19th century, a central challenge in theoretical Physics was describing systems with an enormous number of degrees of freedom (e.g.,
Scope and limitations:
- The core formalism (e.g., phase space, Liouville dynamics) applies universally to classical systems, even far from equilibrium.
- However, the most powerful predictive tools (e.g., Boltzmann distribution,equipartition theorem) assume thermal equilibrium. Beyond equilibrium, approximations (e.g., master equations) are needed, but these still fall under the broader framework of Classical Statistical Mechanics. See the example harmonic oscillator in CSM.
System
In general, you consider a system as a collection of states
Energy
You have an energy function
Volume
You can also have another function
Microstates and macrostates
The points
-
Sometimes these constraints are sharp (hard), meaning they restrict the system to a definite set of microstates. For example,
or more generally an energy window
In both cases the usual assumption is that all compatible microstates are equally probable (principle of equal a priori probability). This is the starting point of the microcanonical description.
-
Other times the constraints are looser (soft), for instance when only the average value of a quantity is fixed. For example, imposing
without forbidding any microstate. In this case, one does not assign uniform probability over a set, but instead determines a probability distribution that satisfies the average constraint. The maximum entropy principle then leads to weighted distributions such as the Boltzmann distribution. This is the basis of the canonical and grand canonical descriptions.
The collection of macroscopic constraints, together with any assumptions used to determine the probability distribution (such as equal a priori probability or maximum entropy), is called a macrostate.
Ensembles
This brings us to the notion of an ensemble: a probability distribution over microstates consistent with the macrostate information.
- Sharp constraints with equal a priori probability give the microcanonical ensemble.
- Average-value constraints (implemented via maximum entropy) give the canonical and grand canonical ensembles.
The motivation is that we cannot control the microscopic coordinatesof a system ( , with large), but only macroscopic variables like mean energy or mean particle number. An ensemble is the probability distribution for the states that corresponds to this macrostate.
Entropy
There are many ensembles compatible with a given macrostate. To choose a single, unbiased ensemble from the many possibilities compatible with a macrostate, we invoke the Principle of Maximum Entropy. This principle states that the best choice for ρ(x) is the one that maximizes the Gibbs entropy (defined below) subject to the known macroscopic constraints. This ensures that we don't assume any information we don't have.
The ensembles have associated a Gibbs/Shannon entropy:
or
in the case of a phase space of a classical Hamiltonian system.
It is interpreted as a measure of our ignorance of the exact microstate of the system. The higher the entropy, the less information we have about where exactly the system is in phase space. A sharply peaked
This functional has the same form as Shannon entropy in information theory and is uniquely characterized by its additivity, continuity, and the fact that it is maximized by uniform distributions under constraints.
Following the principle of maximum entropy we obtain, for example:
- For "narrow" macrostates, such as
, we have a particular kind of ensembles: loosely speaking, a distribution with probability density function
where
- Another important case appears when the macrostate is an average energy constrain
, and maximum entropy is assumed, i.e., thermal equilibrium. This is called the canonical ensemble.
Liouville's equation
The Hamiltonian equations for a particle
correspond to the ideal case of a particle perfectly localized in phase space and we can represent this situation by the "degenerated probability density"
and since
Now, starting from a density
In turn, integrating
Related: