27.10.2013 Views

Fundamental Statistical Mechanics

Fundamental Statistical Mechanics

Fundamental Statistical Mechanics

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Fundamental</strong> <strong>Statistical</strong> <strong>Mechanics</strong><br />

Ref: From Molecular Chaos to Dynamical Chaos, J. R. Dorfman (1995)<br />

6. Boltzmann's Ergodic Hypothesis<br />

6.1 Approach to Equilibrium<br />

6.2 Individual Ergodic Theorem<br />

7. Gibbs' Picture - Mixing systems<br />

7.1 The definition<br />

7.3 Chaos<br />

9. The Baker's transformation<br />

9.1 Transformation and Properties<br />

9.2 A Model Boltzmann Equation<br />

10. Lyapunov exponents for a Map<br />

11. Baker's transformation is Ergodic<br />

11.1 Proof<br />

11.2 Baker's transformation and Irreversibility<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 34


6. Boltzmann's Ergodic Hypothesis<br />

6.1 Approach to Equilibrium<br />

We now begin to try to solve what might well be called the <strong>Fundamental</strong> Problem of <strong>Statistical</strong><br />

<strong>Mechanics</strong>: Mechanical systems - isolated ones, at least - are time reversible and recurrent.<br />

However, we observe that large isolated systems often reach a state of thermodynamic<br />

equilibrium. How do we explain our observations in such a way that there are no contradictions<br />

with the laws of mechanics? Boltzmann proposed a resolution along the following lines:<br />

a) Equilibrium statistical mechanics can be formulated in terms of microcanonical ensemble<br />

averages, by using the invariant measure dµ = dS gradH . The ensemble average of a phase<br />

variable B(Γ), is<br />

B(Γ) mc =<br />

The density of states is<br />

∫<br />

Ω(E) = ∫ dµ<br />

H =E<br />

dµB(Γ)<br />

H = E<br />

dµ<br />

H = E<br />

∫ (6.1)<br />

and the thermodynamic entropy is given by S = k B lnΩ(E).<br />

that<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 35<br />

(6.2)<br />

As an aside, the microcanonical ensemble phase space density is ρ(Γ) = δ(H − E), so<br />

B(Γ) mc =<br />

∫<br />

dΓB(Γ)δ (H − E)<br />

∫ dΓδ(H − E)<br />

(6.3)<br />

If in the laboratory, we were to make very precise measurements of the quantity B(Γ t ), where Γt is the location of the phase point of the system at time t (where for example B(Γ t ) is the force<br />

per unit area on a piston) the values would show wild fluctuations about a more slowly varying<br />

"mean" quantity (as molecules collide with the piston). We might identify the thermodynamic<br />

value with the time average<br />

1<br />

B = B(Γ) t = lim<br />

T →∞ T B(Γ t )dt ∫<br />

T<br />

0<br />

(6.4)


Boltzmann realised that one could identify the microcanonical ensemble average 6.1 with the<br />

infinite time average 6.4, that is<br />

B = B(Γ) mc<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 36<br />

(6.5)<br />

if he made the hypothesis that the trajectory of a typical point on the constant-energy surface<br />

(except for a set of points of zero measure) spends equal time in regions of equal measure. This<br />

hypothesis was called the ergodic hypothesis by Boltzmann and it is of central interest for the<br />

foundations of statistical mechanics.<br />

i<br />

To see how this hypothesis works, subdivide the constant energy surface into a fine grid. The<br />

average of B in each grid region i is B i . Then<br />

T<br />

1<br />

T B(Γ t )dt ∫ ≈<br />

0<br />

∑<br />

i<br />

τ i<br />

T B i<br />

where τi T is the fraction of the time that the trajectories spend in region i between t = 0 and<br />

T . Using the ergodic hypothesis, we can write<br />

so that<br />

τ i<br />

T = µ i<br />

µ(E)<br />

B =<br />

∑<br />

i<br />

µ i<br />

µ(E) Bi = B(Γ) mc<br />

Thus equilibrium statistical mechanics could be justified, for isolated systems, if we could prove<br />

that the ergodic hypothesis is correct for a large class of physical systems.


6.2 Individual Ergodic Theorem<br />

In 1931 Birkhoff proved a very important theorem which allows us to make some of the ideas of<br />

Boltzmann more precise. Although Birkhoff's theorem is still a long way from what is needed in<br />

Boltzmann's picture it does define in a useful way the dynamical properties that a system must<br />

possess in order to have an equilibrium state.<br />

This theorem is concerned with the properties of individual trajectories in phase space. Suppose<br />

we are considering a mechanical system with an invariant measure and suppose that we consider<br />

some phase function B defined on the constant energy surface, satisfying the condition that<br />

∫<br />

dµ B(Γ)


EXAMPLE: Consider the map given by<br />

φ (x) = x + α mod(1)<br />

If α is a rational number, nm where n and m are integers, then the mapping repeats the initial<br />

point after m iterations. A trajectory starting from an initial point will be periodic and will not<br />

spend equal time in regions of equal (Lebesgue) measure, and hence the trajectory is not dense<br />

on the unit interval. In this case the system is not ergodic. But if α is irrational no trajectory is<br />

periodic and the system is ergodic! To see this consider the time average of a phase function on<br />

a trajectory starting at x<br />

1<br />

B (x) = lim<br />

N →∞ N<br />

N −1<br />

∑<br />

n= 0<br />

B(φ n (x))<br />

1<br />

= lim<br />

N→∞ N<br />

N−1<br />

∑<br />

n=0<br />

B(x + nα)<br />

Consider the Fourier series for B(x): B(x) = a j e 2πijx<br />

∞<br />

∑<br />

B(x + nα) = a j e<br />

j =−∞<br />

2πijx+ 2πijnα<br />

Substituting into the time-average gives<br />

1<br />

B (x) = lim<br />

N →∞ N<br />

1<br />

= lim<br />

N →∞ N<br />

N −1<br />

∞<br />

∑ ∑<br />

n= 0 j =−∞<br />

∞<br />

∑<br />

j=−∞<br />

2π ijx+ 2πijnα<br />

a je a j e 2πijx<br />

e 2πijnα<br />

N −1<br />

∑<br />

n= 0<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 38<br />

∞<br />

∑<br />

j =−∞<br />

1<br />

= a0 + lim<br />

N →∞ N<br />

∞<br />

∑<br />

j =−∞<br />

j ≠0<br />

a j e 2πijx<br />

1− e 2πijNα<br />

2 πijα<br />

1 − e<br />

The terms with n ≥ 1 don't survive the limit N →∞, thus B (x) = a 0 for α irrational. In this<br />

case the time average is a constant - independent of the starting point x and the system is<br />

ergodic, where the ensemble average is the average over the circle with the usual Lebesgue<br />

measure,<br />

1<br />

∫<br />

B (x) = dxB(x) = B(x) mc<br />

0


7. Gibbs' Picture - Mixing systems<br />

7.1 The definition<br />

While Boltzmann fixed his attention on the motion of the phase point for a single system and<br />

was led to the concept of ergodicity, Gibbs took another approach to the same problem. Since<br />

one never knows precisely what the initial phase point of a system is, Gibbs decided to consider<br />

the average behaviour of a set of points on the constant energy surface with more or less the<br />

same macroscopic state. Without worrying too much about how such a set might be defined<br />

precisely, let’s consider an initial set of points A . As the set travels though phase space it<br />

changes shape, but its measure stays the same; µ(A) = µ(A t ) . The set gets stretched and folded<br />

and may eventually appear on a coarse enough scale to fill the energy surface uniformly.<br />

However the set At has the same topological structure as the set A and the initial set is not<br />

“forgotten”, in the sense that a time reversal operation on the set At will produce the set A .<br />

There is a nice lecture demonstration apparatus that illustrates this time reversal operation: a drop<br />

of indissoluble ink is added to a container of glycerine. If you stir the glycerine carefully, the<br />

drop will stretch and form a thin line. Eventually it seems to fill the whole space, but if the<br />

stirring is reversed, the initial drop of ink surprisingly reappears.<br />

Gibbs thought that the apparently uniform distribution of the set At on the energy surface was<br />

the key to understanding how mechanically reversible systems could approach an equilibrium<br />

state. To make this idea more precise Gibbs called a system mixing if for each set B<br />

µ(B ∩ At )<br />

lim<br />

t→∞ µ(B)<br />

exists and equals µ(A)<br />

µ(E)<br />

As we will see presently the requirements that a system be mixing is a stronger condition than<br />

ergodicity. However, more can be said about the approach to equilibrium for a mixing system<br />

than an ergodic one.<br />

A<br />

A t<br />

Figure 7.1<br />

To discuss the difference between ergodic and mixing systems we need the notion of metric<br />

indecomposability.<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 39<br />

B


A metrically decomposable system is one for which there exists a subdivision of the constant<br />

energy surface into two regions of non-zero measure, each of which is invariant under the<br />

mechanical flow. That is, a phase point starting out in one region will always stay in that region.<br />

A system is ergodic if and only if it is metrically indecomposable:<br />

1) decomposable → non-ergodic<br />

2) non-ergodic → decomposable<br />

Mixing implies ergodicity. Consider a mixing system and an invariant set A = A t . Then for all<br />

B<br />

µ(A t ∩ B)<br />

lim =<br />

t→∞ µ(B)<br />

µ(A)<br />

µ(E)<br />

or lim<br />

t→∞ µ(A t ∩ B) = µ(A)µ(B)<br />

µ(E)<br />

If we set B = A = A t (since A is an invariant set) then A t ∩ B = A and<br />

µ(A) = µ(A)µ(A)<br />

µ(E)<br />

This equation has two solutions:<br />

1) µ(A) = 0 then one trivial invariant set is a set of zero measure, and the other solution is<br />

2) µ(A) = µ(E); the invariant set is effectively the whole energy surface.<br />

Therefore, if a system is mixing, the only invariant set with positive measure is the constant<br />

energy surface. Any other invariant set must have zero measure. Such sets might be a countable<br />

set of periodic orbits.<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 40


9. The Baker's transformation<br />

9.1 Transformation and Properties<br />

We consider an example which well illustrates the application of ideas from chaotic dynamical<br />

systems to statistical mechanics, the baker's map. We take the phase space to be the unit square<br />

0 ≤ x, y ≤1. The transformation consists of two steps: first the unit square is contracted by a<br />

factor 2 in the y direction and expanded by a factor 2 in the x direction; then the rectangle is cut<br />

in the middle and the right half placed on top of the left half. This doesn't change the volume,<br />

and the transformation is reversible.<br />

To write an express the transformation mathematically we need to distinguish between x < 1 2 and<br />

x ≥ 1 2 , so if (x,y) → ( x ′ , y ′ ) = b(x,y).<br />

For x < 1 2<br />

⎧ x ′ = 2x<br />

⎨<br />

⎩ y ′ = y 2<br />

and for x ≥ 1 2<br />

⎧ x ′ = 2x −1<br />

⎨<br />

⎩ y ′ = (y +1) 2<br />

The inverse of the baker's transformation (x,y) → ( x ′ , y ′ ) = b −1 (x, y)<br />

For y < 1 2<br />

⎧ x ′ = x 2<br />

⎨<br />

⎩ y ′ = 2y<br />

and for y ≥ 1 2<br />

9.2 A Model Boltzmann Equation<br />

⎧ x ′ = (x +1) 2<br />

⎨<br />

⎩ y ′ = 2y −1<br />

It is possible to derive a "Boltzmann equation for the time-reversible baker's transformation and<br />

to show that a H -theorem holds for this equation. Here then is one example where the program<br />

of Boltzmann can be carried out in detail. The price we pay for the simplicity of the model is a<br />

lack of physical motivation. We will have to supply some of that as we go along.<br />

Consider a density-function ρ(x,y) on the unit square, that satisfies a Liouville equation for<br />

discrete time:<br />

where<br />

ρ n (x, y) = ρ n −1 (b −1 (x), b −1 (y))<br />

⎧ ρn−1 (x 2,2y) for y 1 2<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 41


Define a reduced distribution function that depends on x only:<br />

1<br />

∫<br />

W n (x) = dyρ n (x, y)<br />

0<br />

1<br />

2<br />

= ∫ dyρ n−1 (x 2,2y) + ∫ dyρ n −1 ((x +1) 2,2y −1)<br />

0<br />

1<br />

1<br />

2<br />

Change to a variable y ′ = 2y in the first integral and to y ′ = 2y −1 in the second integral:<br />

Wn (x) = 1<br />

2<br />

1<br />

∫<br />

0<br />

d ′ y ρn−1 ( x 2 , ′ y ) +ρ x +1<br />

( n−1 ( 2 , y ′ ) )<br />

= 1<br />

2 W ⎛ x<br />

n−1⎝<br />

2<br />

⎞<br />

⎠ + W ⎛<br />

⎛ x +1⎞<br />

⎞<br />

n−1<br />

⎝<br />

⎝ 2 ⎠ ⎠<br />

This is the model Boltzmann equation that is associated with the baker's transformation. We<br />

notice that the time is discrete rather than continuous, and that we have selected the x coordinate<br />

for some reason that is not yet clear. It is easy to check that if Wn does not depend on x then<br />

Wn remains constant in time. Thus there is an equilibrium distribution W 0 = constant , which<br />

corresponds to a uniform distribution on the unit x interval.<br />

The H -theorem is constructed in the same way as is done for the Boltzmann equation itself. We<br />

define<br />

1<br />

∫<br />

( )<br />

H n = dxW n (x)ln W n (x)<br />

0<br />

Then H develops in time as<br />

1<br />

Hn +1 = dx 1 2 Wn ( x ( 2 ) + Wn ( )ln<br />

∫<br />

0<br />

.<br />

x +1<br />

2 )<br />

1<br />

2 Wn ( x x +1<br />

[ ( 2 ) + Wn ( 2 ) ) ]<br />

as the function F(y) = y ln y is convex, it follows that 1 a +b<br />

2 ( F(a) + F(b) )≥F( 2 ) . Setting<br />

a = Wn ( x x +1<br />

2) and b = Wn ( 2 ) we have:<br />

Hn +1 ≤ 1 2 dx W n ( x 2)ln Wn ( x 1<br />

∫<br />

2)<br />

0<br />

x +1<br />

x +1<br />

( ( )+ Wn ( 2 )ln( W n ( 2 ) ) )<br />

Change to ′<br />

x = x 2 in the first term, and to ′<br />

x = x +1<br />

2 in the second term, we find:<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 42


1<br />

2<br />

( )<br />

Hn +1 ≤ ∫ d x ′ Wn ( x ′ )ln Wn ( x ′ ) + ∫ dxWn ( x ′ )lnW n ( x ′ )<br />

0<br />

That is, we obtain a H -theorem in the form<br />

H n +1 ≤ H n<br />

Note that H stays constant if W is a constant.<br />

1<br />

1<br />

2<br />

( )<br />

Reverting for the moment to a physical system, a dilute gas, we know that the phase space<br />

distribution function is the fundamental distribution which really determines the behaviour of an<br />

ensemble of systems not in equilibrium and that the function which satisfies the Boltzmann<br />

equation is the single particle distribution function, obtained by integrating over the variables of<br />

all but one of the particles. Bogoliubov has argued that one can separate rapidly varying<br />

functions from slowly varying functions, and the physically interesting functions change slowly<br />

with time. The time scales that Bogoliubov thought were relevant in a gas are the duration of<br />

binary collisions, the mean free time between collisions, and the time it takes a particle to travel a<br />

macroscopic distance. Applying Bogoliubov's arguments to the baker's transformation we would<br />

expect that the x variable is slowly varying while the y variable is rapidly varying.<br />

What happens if we integrate the x coordinate rather than the y coordinate?<br />

We need to look at the evolution of density differently<br />

where<br />

ρ n −1 (x, y) = ρ n (b(x), b(y))<br />

⎧ ρn (2x, y 2) forx 1 2<br />

1<br />

∫<br />

V n−1 (y) = dx ρ n−1 (x, y)<br />

0<br />

1<br />

2<br />

= ∫ dx ρn (2x, y 2) + ∫ dx ρn (2x −1,( y +1) 2)<br />

0<br />

1<br />

1<br />

2<br />

= 1<br />

2 d ′ x ρ 1<br />

∫ n ( x ′ , y 2) +<br />

0<br />

1<br />

1<br />

∫ 2<br />

0<br />

= 1 2V ⎛ y<br />

n⎝<br />

2<br />

⎞<br />

⎠ + 1 2 V ⎛ y +1⎞<br />

n⎝<br />

2 ⎠<br />

d ′ x ρn ( x ′ ,(y +1) 2)<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 43


10. Lyapunov exponents for a Map<br />

11. Baker's transformation is Ergodic<br />

11.1 Proof<br />

We now have all the tools needed to prove that the baker's transformation is ergodic. In fact it is<br />

possible to prove much stronger properties of the baker's transformation - it is a Bernoulli<br />

process - which implies that it is mixing. A Bernoulli process is one in which it is possible to<br />

establish some kind of isomorphism between the process and a random Markov process. This is<br />

exactly what we did when we showed that the baker's transformation can be mapped onto a<br />

Bernoulli shifts. Here we will give the proof that the transformation is ergodic since this proof is<br />

simple and very illustrative of methods often employed in more complicated cases.<br />

Consider an infinitesimal neighbourhood of a point (x, y). The vertical line through (x, y) is the<br />

stable manifold of that point, and the future images of nearby points on this line approach the<br />

future images of (x, y) as they travel together. On the horizontal line, the unstable manifold, the<br />

future images of points move away from the future images of (x, y). Under time reversal the role<br />

of the x and y directions are interchanged, the stable manifold becomes the unstable manifold<br />

and vice versa. The invariant measure on the unit square is dµ = dxdy = d x ′ d y ′ = d µ ′ .<br />

Now define forward and backward time averages as<br />

B + (Γ) = lim<br />

n→∞<br />

1<br />

n<br />

B − 1<br />

(Γ) = lim<br />

n→∞ n<br />

n−1<br />

∑<br />

j = 0<br />

n−1<br />

∑<br />

j = 0<br />

B(b j (Γ))<br />

B(b − j (Γ))<br />

Step 1 of the proof is to show that the forward and backward time averages are equal almost<br />

everywhere<br />

B + (Γ) = B − (Γ)<br />

11.2 Baker's transformation and Irreversibility<br />

<strong>Statistical</strong> <strong>Mechanics</strong> Page 44

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!