Elementary Principles in Statistical Mechanics/Chapter XI

CHAPTER XI.

MAXIMUM AND MINIMUM PROPERTIES OF VARIOUS DISTRIBUTIONS IN PHASE.

In the following theorems we suppose, as always, that the systems forming an ensemble are identical in nature and in the values of the external coördinates, which are here regarded as constants.

Theorem I. If an ensemble of systems is so distributed in phase that the index of probability is a function of the energy, the average value of the index is less than for any other distribution in which the distribution in energy is unaltered.

Let us write for the index which is a function of the energy, and for any other which gives the same distribution in energy. It is to be proved that

(419)
where is a function of the energy, and a function of the phase, which are subject to the conditions that
(420)
and that for any value of the energy ()
(421)
Equation (420) expresses the general relations which and must satisfy in order to be indices of any distributions, and (421) expresses the condition that they give the same distribution in energy.

Since is a function of the energy, and may therefore be regarded as a constant within the limits of integration of (421), we may multiply by under the integral sign in both members, which gives

Since this is true within the limits indicated, and for every value of , it will be true if the integrals are taken for all phases. We may therefore cancel the corresponding parts of (419), which gives
(422)
But by (420) this is equivalent to
(423)
Now is a decreasing function of for negative values of , and an increasing function of for positive values of . It vanishes for . The expression is therefore incapable of a negative value, and can have the value 0 only for . The inequality (423) will hold therefore unless for all phases. The theorem is therefore proved.

Theorem II. If an ensemble of systems is canonically distributed in phase, the average index of probability is less than in any other distribution of the ensemble having the same average energy.

For the canonical distribution let the index be , and for another having the same average energy let the index be , where is an arbitrary function of the phase subject only to the limitation involved in the notion of the index, that

(424)
and to that relating to the constant average energy, that
(425)

It is to be proved that

(426)
Now in virtue of the first condition (424) we may cancel the constant term in the parentheses in (426), and in virtue of the second condition (425) we may cancel the term . The proposition to be proved is thus reduced to
which may be written, in virtue of the condition (424),
(427)
In this form its truth is evident for the same reasons which applied to (423).

Theorem III. If is any positive constant, the average value in an ensemble of the expression ( denoting as usual the index of probability and the energy) is less when the ensemble is distributed canonically with modulus , than for any other distribution whatever.

In accordance with our usual notation let us write for the index of the canonical distribution. In any other distribution let the index be .

In the canonical ensemble has the constant value ; in the other ensemble it has the value . The proposition to be proved may therefore be written

(428)
where
(429)
In virtue of this condition, since is constant, the proposition to be proved reduces to
(430)
where the demonstration may be concluded as in the last theorem.

If we should substitute for the energy in the preceding theorems any other function of the phase, the theorems, mutatis mutandis, would still hold. On account of the unique importance of the energy as a function of the phase, the theorems as given are especially worthy of notice. When the case is such that other functions of the phase have important properties relating to statistical equilibrium, as described in Chapter IV,[1] the three following theorems, which are generalizations of the preceding, may be useful. It will be sufficient to give them without demonstration, as the principles involved are in no respect different.

Theorem IV. If an ensemble of systems is so distributed in phase that the index of probability is any function of , , etc., (these letters denoting functions of the phase,) the average value of the index is less than for any other distribution in phase in which the distribution with respect to the functions , , etc. is unchanged.

Theorem V. If an ensemble of systems is so distributed in phase that the index of probability is a linear function of , , etc., (these letters denoting functions of the phase,) the average value of the index is less than for any other distribution in which the functions , , etc. have the same average values.

Theorem VI. The average value in an ensemble of systems of (where denotes as usual the index of probability and any function of the phase) is less when the ensemble is so distributed that is constant than for any other distribution whatever.

Theorem VII. If a system which in its different phases constitutes an ensemble consists of two parts, and we consider the average index of probability for the whole system, and also the average indices for each of the parts taken separately, the sum of the average indices for the parts will be either less than the average index for the whole system, or equal to it, but cannot be greater. The limiting case of equality occurs when the distribution in phase of each part is independent of that of the other, and only in this case.

Let the coördinates and momenta of the whole system be , of which relate to one part of the system, and to the other. If the index of probability for the whole system is denoted by , the probability that the phase of an unspecified system lies within any given limits is expressed by the integral

(431)
taken for those limits. If we set
(432)
where the integrations cover all phases of the second system, and
(433)
where the integrations cover all phases of the first system, the integral (431) will reduce to the form
(434)
when the limits can be expressed in terms of the coördinates and momenta of the first part of the system. The same integral will reduce to
(435)
when the limits can be expressed in terms of the coördinates and momenta of the second part of the system. It is evident that and are the indices of probability for the two parts of the system taken separately.

The main proposition to be proved may be written

(436)
where the first integral is to be taken over all phases of the first part of the system, the second integral over all phases of the second part of the system, and the last integral over all phases of the whole system. Now we have
(437)
(438)
and
(439)
where the limits cover in each case all the phases to which the variables relate. The two last equations, which are in themselves evident, may be derived by partial integration from the first.

It appears from the definitions of and that (436) may also be written

(440)
or
where the integrations cover all phases. Adding the equation
(441)
which we get by multiplying (438) and (439), and subtracting (437), we have for the proposition to be proved
(442)
Let
(443)
The main proposition to be proved may be written
(444)
This is evidently true since the quantity in the parenthesis is incapable of a negative value.[2] Moreover the sign can hold only when the quantity in the parenthesis vanishes for all phases, i. e., when for all phases. This makes for all phases, which is the analytical condition which expresses that the distributions in phase of the two parts of the system are independent.

Theorem VIII. If two or more ensembles of systems which are identical in nature, but may be distributed differently in phase, are united to form a single ensemble, so that the probability-coefficient of the resulting ensemble is a linear function of the probability-coefficients of the original ensembles, the average index of probability of the resulting ensemble cannot be greater than the same linear function of the average indices of the original ensembles. It can be equal to it only when the original ensembles are similarly distributed in phase.

Let , , etc. be the probability-coefficients of the original ensembles, and that of the ensemble formed by combining them; and let , , etc. be the numbers of systems in the original ensembles. It is evident that we shall have

(445)
where
(446)
The main proposition to be proved is that
(447)
or
(448)

If we set

will be positive, except when it vanishes for . To prove this, we may regard and as any positive quantities. Then
Since and vanish for , and the second differential coefficient is always positive, must be positive except when . Therefore, if , etc. have similar definitions,
(449)
But since
and
(450)
This proves (448), and shows that the sign will hold only when
for all phases, i. e., only when the distribution in phase of the original ensembles are all identical.

Theorem IX. A uniform distribution of a given number of systems within given limits of phase gives a less average index of probability of phase than any other distribution.

Let be the constant index of the uniform distribution, and the index of some other distribution. Since the number of systems within the given limits is the same in the two distributions we have

(451)
where the integrations, like those which follow, are to be taken within the given limits. The proposition to be proved may be written
(452)
or, since is constant,
(453)
In (451) also we may cancel the constant factor , and multiply by the constant factor . This gives
The subtraction of this equation will not alter the inequality to be proved, which may therefore be written
or
(454)
Since the parenthesis in this expression represents a positive value, except when it vanishes for , the integral will be positive unless vanishes everywhere within the limits, which would make the difference of the two distributions vanish. The theorem is therefore proved.

  1. See pages 37-41.
  2. See Theorem I, where this is proved of a similar expression.