Processing math: 100%
Skip to main content
Library homepage
 

Text Color

Text Size

 

Margin Size

 

Font Type

Enable Dyslexic Font
Physics LibreTexts

2.1: Statistical ensemble and probability

( \newcommand{\kernel}{\mathrm{null}\,}\)

As has been already discussed in Sec. 1.1, statistical physics deals with situations when either unknown initial conditions, or system’s complexity, or the laws of its motion (as in the case of quantum mechanics) do not allow a definite prediction of measurement results. The main formalism for the analysis of such systems is the probability theory, so let me start with a very brief review of its basic concepts, using an informal “physical” language – less rigorous but (hopefully) more transparent than standard mathematical treatments,1 and quite sufficient for our purposes.

Consider N>>1 independent similar experiments carried out with apparently similar systems (i.e. systems with identical macroscopic parameters such as volume, pressure, etc.), but still giving, by any of the reasons listed above, different results of measurements. Such a collection of experiments, together with a fixed method of result processing, is a good example of a statistical ensemble. Let us start from the case when the experiments may have M different discrete outcomes, and the number of experiments giving the corresponding different results is N1,N2,...,NM, so that

Mm=1Nm=N.

The probability of each outcome, for the given statistical ensemble, is then defined as

Probability:

WmlimNNmN.

Though this definition is so close to our everyday experience that it is almost self-evident, a few remarks may still be relevant.

First, the probabilities Wm depend on the exact statistical ensemble they are defined for, notably including the method of result processing. As the simplest example, consider throwing the standard cubic-shaped dice many times. For the ensemble of all thrown and counted dice, the probability of each outcome (say, “1”) is 1/6. However, nothing prevents us from defining another statistical ensemble of dice-throwing experiments in which all outcomes “1” are discounted. Evidently, the probability of finding outcomes “1” in this modified (but legitimate) ensemble is 0, while for all other five outcomes (“2” to “6”), it is 1/5 rather than 1/6.

NmWmN,

with the relative deviations decreasing as 1/Nm1/2, i.e. as 1/N1/2.

Now let me list those properties of probabilities that we will immediately need. First, dividing both sides of Equation (???) by N and following the limit N, we get the well-known normalization condition

Mm=1Wm=1;

just remember that it is true only if each experiment definitely yields one of the outcomes N1,N2,...,NM.

Second, if we have an additive function of the results,

f=1NMm=1Nmfm,

where fm are some definite (deterministic) coefficients, the statistical average (also called the expectation value) of the function is naturally defined as

flimN1NMm=1Nmfm,

so that using Equation (???) we get

Expectation value via probabilities:

f=Mm=1Wmfm.

Notice that Equation (???) may be considered as the particular form of this general result, when all fm=1.

Next, the spectrum of possible experimental outcomes is frequently continuous for all practical purposes. (Think, for example, about the set of positions of the marks left by bullets fired into a target from afar.) The above formulas may be readily generalized to this case; let us start from the simplest situation when all different outcomes may be described by just one continuous scalar variable q – which replaces the discrete index m in Eqs. (???)-(???). The basic relation for this case is the self-evident fact that the probability dW of having an outcome within a small interval dq near some point q is proportional to the magnitude of that interval:

dW=w(q)dq,

where w(q) is some function of q, which does not depend on dq. This function is called probability density. Now all the above formulas may be recast by replacing the probabilities Wm with the products (???), and the summation over m, with the integration over q. In particular, instead of Equation (???) the normalization condition now becomes

w(q)dq=1,

where the integration should be extended over the whole range of possible values of q. Similarly, instead of the discrete values fm participating in Equation (???), it is natural to consider a function f(q). Then instead of Equation (???), the expectation value of the function may be calculated as

Expectation value via probability density:

f=w(q)f(q)dq.

It is also straightforward to generalize these formulas to the case of more variables. For example, the state of a classical particle with three degrees of freedom may be fully described by the probability density w defined in the 6D space of its generalized radius-vector q and momentum p. As a result, the expectation value of a function of these variables may be expressed as a 6D integral

f=w(q,p)f(q,p)d3qd3p.

Some systems considered in this course consist of components whose quantum properties cannot be ignored, so let us discuss how f should be calculated in this case. If by fm we mean measurement results, then Equation (???) (and its generalizations) remains valid, but since these numbers themselves may be affected by the intrinsic quantum-mechanical uncertainty, it may make sense to have a bit deeper look into this situation. Quantum mechanics tells us4 that the most general expression for the expectation value of an observable f in a certain ensemble of macroscopically similar systems is

f=m,mWmmfmmTr(Wf).

Here fmm are the matrix elements of the quantum-mechanical operator ˆf corresponding to the observable f, in a full basis of orthonormal states m,

fmm=m|ˆf|m,

while the coefficients Wmm are the elements of the so-called density matrix W, which represents, in the same basis, the density operator ˆW describing properties of this ensemble. Equation (???) is evidently more general than Equation (???), and is reduced to it only if the density matrix is diagonal:

Wmm=Wmδmm

(where δmm is the Kronecker symbol), when the diagonal elements Wm play the role of probabilities of the corresponding states.

Thus formally, the largest difference between the quantum and classical description is the presence, in Equation (???), of the off-diagonal elements of the density matrix. They have the largest values in the pure (also called “coherent”) ensemble, in which the state of the system may be described with state vectors, e.g., the ket-vector

|α=mαm|m

where αm are some (generally, complex) coefficients. In this case, the density matrix elements are merely

Wmm=αmαm,

so that the off-diagonal elements are of the same order as the diagonal elements. For example, in the very important particular case of a two-level system, the pure-state density matrix is

W=(α1α1α1α2α2α1α2α2),

so that the product of its off-diagonal components is as large as that of the diagonal components.

In the most important basis of stationary states, i.e. the eigenstates of the system’s time independent Hamiltonian, the coefficients αm oscillate in time as5

αm(t)=αm(0)exp{iEmt}|αm|exp{iEmt+iφm},

where Em are the corresponding eigenenergies, and φm are constant phase shifts. This means that while the diagonal terms of the density matrix (???) remain constant, its off-diagonal components are oscillating functions of time:

Wmm=αmαm=|αmαm|exp{iEmEmt}exp{i(φmφm)}

Due to the extreme smallness of the Planck constant (on the human scale of things), minuscule random perturbations of eigenenergies are equivalent to substantial random changes of the phase multipliers, so that the time average of any off-diagonal matrix element tends to zero. Moreover, even if our statistical ensemble consists of systems with exactly the same Em, but different values φm (which are typically hard to control at the initial preparation of the system), the average values of all Wmm (with mm) vanish again.

This is why, besides some very special cases, typical statistical ensembles of quantum particles are far from being pure, and in most cases (certainly including the thermodynamic equilibrium), a good approximation for their description is given by the opposite limit of the so-called classical mixture, in which all off-diagonal matrix elements of the density matrix equal zero, and its diagonal elements Wmm are merely the probabilities Wm of the corresponding eigenstates. In this case, for the observables compatible with energy, Equation (???) is reduced to Equation (???), with fm being the eigenvalues of the variable f, so that we may base our further discussion on this key relation and its continuous extensions (???)-(???).


This page titled 2.1: Statistical ensemble and probability is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by Konstantin K. Likharev via source content that was edited to the style and standards of the LibreTexts platform.

Support Center

How can we help?

2: Principles of Physical Statistics
2.2: Microcanonical ensemble and distribution