Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Termodinámica Estadística

Download as pdf or txt
Download as pdf or txt
You are on page 1of 97

Lecture Notes

Advanced Physical Chemistry

Statistical Thermodynamics
Gunnar Jeschke
Copyright 2015 Gunnar Jeschke

Title image: Billard von No-w-ay in collaboration with H. Caps - Eigenes Werk.
Lizenziert unter GFDL ber Wikimedia Commons -
https://commons.wikimedia.org/wiki/File:Billard.JPG
Chapter 2 Word cloud: http://de.123rf.com/profile_radiantskies
Chapter 3 Dice image: http://de.123rf.com/profile_whitetag
Chapter 4 Matryoshka image: http://de.123rf.com/profile_mikewaters
Chapter 5 Word cloud: http://de.123rf.com/profile_radiantskies
Chapter 7 Dumbbell image: http://de.123rf.com/profile_filipobr
Chapter 8 Spaghetti image: http://de.123rf.com/profile_winterbee

Published by Gunnar Jeschke

http://www.epr.ethz.ch

Licensed under the Creative Commons Attribution-NonCommercial 3.0 Unported License (the
License). You may not use this file except in compliance with the License. You may obtain a
copy of the License at http://creativecommons.org/licenses/by-nc/3.0.

Design and layout of the lecture notes are based on the Legrand Orange Book available at
http://latextemplates.com/template/the-legrand-orange-book.

First printing, September 2015


Contents

1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.1 General Remarks 7
1.2 Suggested Reading 8
1.3 Acknowledgment 9

2 Basics of Statistical Mechanics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11


2.1 Basic Assumptions of Statistical Thermodynamics 11
2.1.1 Thermodynamics Based on Statistical Mechanics . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.1.2 The Markovian Postulate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.2 Phase space 13


2.2.1 Hamiltonian Equations of Motion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.2.2 The Liouville Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
2.2.3 Quantum Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

2.3 Statistical Mechanics Based on Postulates 15


2.3.1 The Penrose Postulates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.3.2 Implications of the Penrose Postulates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

3 Probability Theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.1 Discrete Probability Theory 17
3.1.1 Discrete Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
3.1.2 Multiple Discrete Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.1.3 Functions of Discrete Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
3.1.4 Discrete Probability Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
3.1.5 Probability Distribution of a Sum of Random Numbers . . . . . . . . . . . . . . . . . . . . . . . . . 21
3.1.6 Binomial Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.7 Stirlings Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.2 Continuous Probability Theory 24
3.2.1 Probability Density . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.2.2 Selective Integration of Probability Densities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.2.3 Sum of Two Continuous Random Numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

4 Classical Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4.1 Statistical Ensembles 31
4.1.1 Concept of an Ensemble . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4.1.2 Ergodicity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

4.2 Microcanonical Ensemble 32


4.3 Canonical Ensemble 33
4.3.1 Boltzmann Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
4.3.2 Equipartition Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.3.3 Internal Energy and Heat Capacity of the Canonical Ensemble . . . . . . . . . . . . . . . . . . . . 38

4.4 Grand Canonical Ensemble 39

5 Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1 Swendsens Postulates of Thermodynamics 41
5.1.1 Cautionary Remarks on Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1.2 Swendsens Postulates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1.3 Entropy in Phenomenological Thermodynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
5.1.4 Boltzmanns Entropy Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

5.2 The Relation of State Functions to the Partition Function 44


5.2.1 Entropy and the Partition Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
5.2.2 Helmholtz Free Energy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
5.2.3 Gibbs Free Energy, Enthalpy, and Pressure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

5.3 Irreversibility 47
5.3.1 Historical Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
5.3.2 Irreversibility as an Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

5.4 Entropy and Information 48


5.4.1 Gibbs Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
5.4.2 Von Neumann Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
5.4.3 Shannon Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

6 Quantum Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
6.1 Quantum Canonical Ensemble 51
6.1.1 Density Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
6.1.2 Quantum Partition Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

6.2 Quantum and Classical Statistics 54


6.2.1 Types of Permutation Symmetry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
6.2.2 Bose-Einstein Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
6.2.3 Fermi-Dirac Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
6.2.4 Maxwell-Boltzmann Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
6.3 Simple Quantum Systems 58
6.3.1 Spin S = 1/2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
6.3.2 Harmonic Oscillator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
6.3.3 Einstein and Debye Models of a Crystal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61

7 Partition Functions of Gases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65


7.1 Separation of contributions 65
7.1.1 Collective Degrees of Freedom . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
7.1.2 Factorization of Energy Modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

7.2 Translational Partition Function 67


7.2.1 Density of States of an Ideal Gas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
7.2.2 Partition Function and Accessible States . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68

7.3 Nuclear Spin Partition Function 69


7.3.1 High-Temperature Limit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7.3.2 Symmetry Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70

7.4 Rotational Partition Function 71


7.4.1 Rigid Rotor Assumption and Rotamers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
7.4.2 Accessible States and Symmetry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72

7.5 Vibrational Partition Function 74


7.5.1 The Harmonic Oscillator Extended . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
7.5.2 Vibrational Contributions to U , CV , and S . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75

7.6 Electronic Partition Function 75


7.7 Equilibrium Constant for Gas Reactions 77

8 Macromolecules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.1 Thermodynamics of Mixing 81
8.1.1 Entropy of Binary Mixing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.1.2 Energy of Binary Mixing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83

8.2 Entropic Elasticity 85


8.2.1 Ideal Chain Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
8.2.2 Random Walk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
8.2.3 Conformational Entropy and Free Energy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Books 93
Articles 94
Web Pages 94

Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
General Remarks
Suggested Reading
Acknowledgment

1 Introduction

Today, Mr. K. complained, Scores of people claim in public that they can type sizeable books
entirely on their own, and this is universally accepted. The Chinese philosopher Chuan-Tzu,
when already at his prime age, brushed a tome of one hundred thousand words with nine out of
ten being citations. Such books cannot be written anymore, since wit is missing. Hence, only by
the tools of a single man thoughts are produced, whereas he feels lazy if their number is low.
Indeed, no thought can be reused and no expression be cited. How little they all need for their
doings! A pen and some paper is all they have to show! And without any help, with only the puny
material that a single man can carry on his arms, they set up their huts! They dont know of
larger buildings than a loner can raise.

Freely translated from: Bertolt Brecht, Geschichten vom Herrn Keuner

1.1 General Remarks


The field of Statistical Thermodynamics is probably the branch of physical chemistry whose
coverage in textbooks is most diverse. A canonical way of teaching this subject still appears to
be missing, which is partially due to the fact that practitioners have not completely agreed on
interpretation of the probabilistic character, on the postulates, and on the way how the mathematical
apparatus is derived from a set of postulates. While this may appear troublesome, actually
there exists rarely any problem in applying statistical thermodynamics1. Accordingly, textbooks
usually ignore the more problematic aspects and try to give reasons why the interpretation and
formulation used by the respective author should be the preferred one. This being an advanced
lecture course, we shall not do so, but we shall still present an apparatus that is ready-made for
application.
The basic idea of statistical thermodynamics is simple: On the one hand we have Newtonian
and quantum mechanics and we know that molecules should adhere to it, and on the other
hand we know that systems consisting of many molecules can be adequately described by
phenomenological (or classical) thermodynamics. Now lets try to derive the latter theory from
the former one. Some care will have to be taken for systems that are subject to quantum statistics,
but we might expect that straightforward application of probability theory will provide the
1The problematic cases are mostly those where the number of particles is too small for the approximations made
in the statistical approach.
8 Introduction

required connection. Chapter 2 will discuss this basic idea in some more detail and will present a
set of postulates due to Oliver Penrose [Pen70]. The discussion of these postulates clarifies what
the remaining mathematical problem is and how we avoid it in applications.
In this course we do not assume that students are already familiar with probability theory,
rather we will introduce its most important concepts in Chapter 3. We do assume that the concepts
of phenomenological thermodynamics are known, although we shall shortly explain them on
first use in these lecture notes. The most important new concept in this course is the one of an
ensemble description, which will be introduced in Chapter 4 first only for classical particles.
This will set the stage for discussing the concepts of irreversibility and entropy in Chapter 5. We
will complete the foundations part with a discussion of quantum ensembles in Chapter 6. This
Chapter will also make the transition to applications, by treating first the harmonic oscillator and
second the Einstein model of a crystal with the apparatus that we command at that point.
We shall then illustrate the relation to phenomenological thermodynamics by discussing
the partition functions of gases and by computing thermodynamical state functions from these
partition functions in Chapter 7. The final Chapter 8 will shortly discuss the consequences of
statistical thermodynamics for macromolecular systems and introduce the concepts of lattice
models, random walks, and entropic elasticity.
The time available for this course does not permit to treat all aspects of statistical thermo-
dynamics and statistical mechanics that are important in physical chemistry, chemical physics,
polymer physics, and biophysics, let alone in solid-state physics. The most important omissions
are probably kinetic aspects of chemical reactions, which are treated in detail in a lecture course
on Advanced Kinetics, and the topic of phase transitions, including the famous Ising chain model.
We believe that the foundations laid in the present course will allow students to understand these
topics from reading in the textbooks listed in the following Section.

1.2 Suggested Reading


General textbooks on physical chemistry, such as [AP13; ER06] provide some overview of the
most important concepts of statistical thermodynamics as well as some of the key formulas, but
they are not quite on the level of this advanced course. If you already own these books, it might
still be useful to read what they write on the topic. If you have the choice, Engel/Reid [ER06] is
better on this topic than Atkins [AP13]. The best introduction in a general physical chemistry
textbook can be found in the German book by Wedler and Freund [WF12].
A rather good and modern introduction at an advanced level has been published in English
language by Swendsen [Swe12]. Swendsen introduces statistical thermodynamics together with
phenomenological thermodynamics and covers more examples than we can treat in this course.
He does not introduce some concepts that are widely used in the field, because he dislikes them.
In this course we do introduce these concepts and discuss the problems associated with them.
A modern German-language introduction is the one by Schwabl [Sch06], which caters more
to the physicist than to the physical chemist. Schwabl is stronger on phase transitions and dynamic
phenomena, but probably harder to read than Swendsen, even if German is your native language.
For Chapter 7, but only for this Chapter, Maczeks book from the Oxford Chemistry Primers
series [Mac98] can be quite useful. Several topics that are not or only superficially covered in my
lecture notes are treated in the notes by Cohen from Ben Gurion University [Coh11], which are
strongly focused on quantum applications. Finally, I want to mention Penroses book [Pen70],
which is certainly not an introductory textbook and may be most appealing to the strongly
mathematically or philosophically inclined. If you look for guidance on applying statistical
thermodynamics to real systems this book is certainly deficient, but from an epistemological
point of view it is probably the best one.
1.3 Acknowledgment 9

For many of the central concepts I have looked up (English) Wikipedia articles and have
found that these articles are, on average, of rather good quality. They do differ quite strongly
from each other in style and notation. When using only Wikipedia or other internet resources it
is difficult to fit the pieces of information together. If, on the other hand, you already do have a
basic level of understanding, but some difficulties with a particular concept, such sources may
provide just the missing piece of information. The NIST guide for computing thermodynamical
state functions from the results of ab initio computations is a particularly good example for a
useful internet resource [Iri98].

1.3 Acknowledgment
I am grateful to M. Schfer, U. Hollenstein, and F. Merkt for making their lecture notes for this
course available and to Takuya Segawa for thorough proofreading of the first manuscript of these
notes. All remaining errors are my own sole responsibility.
Basic Assumptions of Statistical
Thermodynamics
Thermodynamics Based on Statistical Mechanics
The Markovian Postulate
Phase space
Hamiltonian Equations of Motion
The Liouville Equation
Quantum Systems
Statistical Mechanics Based on Postulates
The Penrose Postulates
Implications of the Penrose Postulates

2 Basics of Statistical Mechanics

2.1 Basic Assumptions of Statistical Thermodynamics


2.1.1 Thermodynamics Based on Statistical Mechanics
Phenomenological thermodynamics describes relations between observable quantities that
characterize macroscopic material objects. We know that these objects consist of a large number
of small particles, molecules or atoms, and, for all we know, these small particles adhere to
the laws of quantum mechanics and often in good approximation to the laws of Newtonian
mechanics. Statistical mechanics is the theory that explains macroscopic properties, not only
thermodynamic state functions, by applying probability theory to the mechanic equations of
motion for a large ensemble of systems of particles. In this lecture course we are concerned with
the part of statistical mechanics that relates to phenomenological thermodynamics.
In spite of its name, phenomenological (equilibrium) thermodynamics is essentially a static
theory that provides an observational, macroscopic description of matter. The underlying
mechanical description is dynamical and microscopic, but it is observational only for systems
consisting of a small number of particles. To see this, we consider a system of N identical
classical point particles that adhere to Newtons equations of motion.

Concept 2.1.1 Newtonian equations of motion. With particle mass m, Cartesian coordinates
qi (i = 1, 2, . . . , 3N ) and velocity coordinates qi , a system of N identical classical point
particles evolves by

d2 qi
m 2
= V (q1 , . . . , q3N ) , (2.1)
dt qi

where V (q1 , . . . , q3N ) is the potential energy function.

Notation 2.1. The dynamical state or microstate of the system at any instant is defined by the
6N Cartesian and velocity coordinates, which span the dynamical space of the system. The
curve of the system in dynamical space is called a trajectory.

The concept extends easily to atoms with different masses mi . If we could, at any instant,
precisely measure all 6N dynamical coordinates, i.e., spatial coordinates and velocities, we could
12 Basics of Statistical Mechanics

precisely predict the future trajectory. The system as described by the Newtonian equations of
motions behaves deterministically.
For any system that humans can see and handle directly, i.e., without complicated technical
devices, the number N of particles is too large (at least of the order of 1018 ) for such complete
measurements to be possible. Furthermore, for such large systems even tiny measurement errors
would make the trajectory prediction useless after a rather short time. In fact, atoms are quantum
objects and the measurements are subject to the Heisenberg uncertainty principle, and even the
small uncertainty introduced by that would make a deterministic description futile.
We can only hope for a theory that describes what we can observe. The number of
observational states or macrostates that can be distinguished by the observer is much smaller than
the number of dynamical states. Two classical systems in the same dynamical state are necessarily
also in the same observational state, but the converse is not generally true. Furthermore, the
observational state also evolves with time, but we have no equations of motion for this state (but
see Section 2.2.2). In fact we cannot have deterministic equations of motion for the observational
state of an individual system, precisely because the same observational state may correspond to
different dynamical states that will follow different trajectories.
Still we can make predictions, only these predictions are necessarily statistical in nature. If
we consider a large ensemble of identical systems in the same observational state we can even
make fairly precise predictions about the outcome. Penrose [Pen70] gives the example of a
women at a time when ultrasound diagnosis can detect pregnancy, but not sex of the foetus. The
observational state is pregnancy, the two possible dynamical states are on path to a boy or girl.
We have no idea what will happen in the individual case, but if the same diagnosis is performed
on a million of women, we know that about 51-52% will give birth to a boy.
How then can we derive stable predictions for an ensemble of systems of molecules? We
need to consider probabilities of the outcome and these probabilities will become exact numbers
in the limit where the number N of particles (or molecules) tends to infinity. The theory required
for computing such probabilities will be treated in Chapter 3.

R Our current usage of the term ensemble is loose. We will devote the whole Chapter 4 to
clarifying what types of ensembles we use in computations and why.

2.1.2 The Markovian Postulate


There are different ways for defining and interpreting probabilities. For abstract discussions and
mathematical derivations the most convenient definition is the one of physical or frequentist
probability.
Given a reproducible trial T of which A is one of the
Definition 2.1.1 Physical probability.
possible outcomes, the physical probability P of the outcome A is defined as

n(A, N , T )
P (A|T ) = lim (2.2)
N N
where n(A, N , T ) is the number of times the outcome A is observed in the first N trials.

A trial T conforming to this definition is statistically regular, i.e., the limit exists and is the
same for all infinite series of the same trial. If the physical probability is assumed to be a stable
property of the system under study, it can be measured with some experimental error. This
experimental error has two contributions: (i) the actual error of the measurement of the quantity
A and (ii) the deviation of the experimental frequency of observing A from the limit defined in
Eq. (2.2). Contribution (ii) arises from the experimental number of trials N not being infinite.
2.2 Phase space 13

We need some criterion that tells us whether T is statistically regular. For this we split the
trial into a preparation period, an evolution period, and the observation itself. The evolution
period is a waiting time during which the system is under controlled conditions. Together with
the preparation period it needs to fulfill the Markovian postulate.

Concept 2.1.2 Markovian postulate. A trial T that invariably ends up in the observational
state O of the system after the preparation stage is called statistically regular. The start of the
evolution period is assigned a time t = 0.

Note that the system can be in different observational states at the time of observation;
otherwise the postulate would correspond to a trivial experiment. The Markovian postulate is
related to the concept of a Markovian chain of events. In such a chain the outcome of the next
event depends only on the current state of the system, but not on states that were encountered
earlier in the chain. Processes that lead to a Markovian chain of events can thus be considered as
memoryless.

2.2 Phase space

2.2.1 Hamiltonian Equations of Motion

The Newtonian equations of motion (2.1) are very convenient for atomistic molecular dynamics
(MD) computations. Statistical analysis of trajectories encountered during such MD simulations
can be analyzed in terms of thermodynamic quantities, such as free energy. However, for
analyzing evolution of the system in terms of spectroscopic properties, the Newtonian description
is very inconvenient. Since spectroscopic measurements can provide the most stringent tests of
theory, we shall use the Hamiltonian formulation of mechanics in the following. This formulation
is particularly convenient for molecules that also have rotational degrees of freedom. For that,
we replace the velocity coordinates by momentum coordinates pj = mj qj , where index j runs
over all atoms. Furthermore, we assume M identical molecules, with each of them having f
degrees of freedom, so that the total number of degrees of freedom is F = f M . Such as system
can be described by 2F differential equations

Concept 2.2.1 Hamiltonian equations of motion. With the single-molecule Hamiltonian


H(pi , qi ) the equations of motion for M non-interacting identical molecules with f degrees
of freedom for each molecule read
dqi H (pi , qi )
= (2.3)
dt pi
dpi H (pi , qi )
= , (2.4)
dt qi
where i = 1 . . . M . Each of the dynamical variables qi and pi is a vector of length f . The
2f M dynamical variables span the phase space.

Definition 2.2.1 Phase space and state space.Phase space is the space where microstates
of a system reside. Sometimes the term is used only for problems that can be described
in spatial and momentum coordinates, sometimes for all problems where some type of a
Hamiltonian equation of motion applies. Sometimes the term state space is used for the
space of microstates in problems that cannot be described by (only) spatial and momentum
coordinates.
14 Basics of Statistical Mechanics

If the molecule is just a single atom, we have only f = 3 translational degrees of freedom
and the Hamiltonian is given by
1
p2x,i + p2y,i + p2z,i ,

H (pi , qi ) = (2.5)
2m
describing translation. For molecules with n atoms, three of the f = 3n degrees of freedom are
translational, two or three are rotational for linear and non-linear molecules, respectively, and the
remaining 3n 5 or 3n 6 degrees of freedom are vibrational.

2.2.2 The Liouville Equation


Our observations do not allow us to specify phase space trajectories, i.e. the trajectory of
microstates for a single system. Instead, we consider an ensemble of identical systems that
all represent the same (observational) macrostate O but may be in different microstates. At a
given time we can characterize such an ensemble by a probability density (p, q, t) in phase
space, where p and q are the vectors of all momentum and spatial coordinates in the system,
respectively. We are interested in an equation of motion for this probability density , which
corresponds to the full knowledge that we have on the system. This equation can be derived from
an integral representation of and the Hamiltonian equations of motion [Sch06].

The probability density (p, q, t) in phase space evolves


Concept 2.2.2 Liouville Equation.
in time according to
 
X H H
= . (2.6)
t pi qi qi pi
i

With the Poisson brackets


X  u v u v

{u, v} = . (2.7)
pi qi qi pi
i

this Liouville equation can be expressed as


= {H, } . (2.8)
t
For the probability density along a phase space trajectory, i.e., along a trajectory that is taken
by microstates, we find

d
(q(t), p(t), t) = 0 . (2.9)
dt
If we consider a uniformly distributed number dN of ensemble members in a volume element
d0 in phase space at time t = 0 and ask about the volume element d in which these ensemble
members are distributed at a later time, we find

d = d0 . (2.10)

This is the Liouville theorem of mechanics.

2.2.3 Quantum Systems


Hamiltonian mechanics can be applied to quantum systems, with the Hamiltonian equations of
motion being replaced by the time-dependent Schrdinger equation. The probability density in
2.3 Statistical Mechanics Based on Postulates 15

phase space is replaced by the density operator b and the Liouville equation by the Liouville-von-
Neumann equation
b i hb i
= H, b . (2.11)
t ~
In quantum mechanics, observables are represented by operators A.b The expectation value of
an observable can be computed from the density operator that represents the distribution of the
ensemble in phase space,
 
hAi
b = Trace bA b . (2.12)

We note that the Heisenberg uncertainty relation does not introduce an additional complication
in statistical mechanics. Determinism had been lost before and the statistical character of the
measurement on an individual system is unproblematic, as we seek only statistical predictions
for a large ensemble. In the limit of an infinite ensemble, N , there is no uncertainty
and the expectation values of incompatible observables are well defined and can be measured
simultaneously. Such an infinitely large system is not perturbed by the act of observing it. The
only difference between the description of classical and quantum systems arises from their
statistical behavior on permutation of the coordinates of two particles, see Section 6.2.

2.3 Statistical Mechanics Based on Postulates


2.3.1 The Penrose Postulates
Penrose [Pen70] has made the attempt to strictly specify what results can be expected from
statistical mechanics if the theory is based on a small number of plausible postulates.
1. Macroscopic physical systems are composed of molecules that obey classical or quantum
mechanical equations of motion (dynamical description of matter).
2. An observation on such a macroscopic system can be idealized as an instantaneous,
simultaneous measurement of a set of dynamical variables, each of which takes the values
1 or 0 only (observational description of matter).
3. A measurement on the system has no influence whatsoever on the outcome of a later
measurement on the same system (compatibility).
4. The Markovian postulate. (Concept 2.1.2)
5. Apart from the Bose and Fermi symmetry conditions for quantum systems, the whole
phase space can, in principle, be accessed by the system (accessibility).
After the discussion above, only the second of these postulates may not immediately appear
plausible. In the digital world of today it appears natural enough: Measurements have resolution
limits and their results are finally represented in a computer by binary numbers, which can be
taken to be the dynamical variables in this postulate.

2.3.2 Implications of the Penrose Postulates


Entropy is one of the central quantities of thermodynamics, as it tells in which direction a
spontaneous process in an isolated system will proceed. For closed systems that can exchange
heat and work with their environment, such predictions on spontaneous processes are based
on free energy, of which the entropy contribution is usually an important part. To keep such
considerations consistent, entropy must have two fundamental properties
1. If the system does not exchange energy with its environment, its entropy cannot decrease.
(non-decrease).
2. The entropy of two systems considered together is the sum of their separate entropies.
(additivity).
16 Basics of Statistical Mechanics

Based on the Penrose postulates it can be shown [Pen70] that the definition of Boltzmann entropy
(Chapter 5) ensures both properties, but that statistical expressions for entropy ensure only the
non-decrease property, not in general the additivity property. This appears to leave us in an
inconvenient situation. However, it can also be shown that for large systems, in the sense that the
number of macrostates is much smaller than the number of microstates, the term that quantifies
non-additivity is negligibly small compared to the total entropy [Pen70]. The problem is thus
rather a mathematical beauty spot than a serious difficulty in application of the theory.
Discrete Probability Theory
Discrete Random Variables
Multiple Discrete Random Variables
Functions of Discrete Random Variables
Discrete Probability Distributions
Probability Distribution of a Sum of Random
Numbers
Binomial Distribution
Stirlings Formula
Continuous Probability Theory
Probability Density
Selective Integration of Probability Densities
Sum of Two Continuous Random Numbers

3 Probability Theory

3.1 Discrete Probability Theory


3.1.1 Discrete Random Variables
Consider a trial T where the observation is a measurement of the z component ~mS of spin
angular momentum of a spin S = 5/2. There are just six possible outcomes (events) that can
be labeled with the magnetic spin quantum number mS or indexed by integer numbers 1, 2, . . .
6. In general, the probabilities of the six possible events will differ from each other. They will
depend on preparation and may depend on evolution time before the observation. To describe
such situations, we define a set of elementary events

A = {aj } , (3.1)

where in our example index j runs from 1 to 6, whereas in general it runs from 1 to the number
NA of possible events. Each of the events is assigned a probability 0 P (aj ) 1. Impossible
events (for a given preparation) have probability zero and a certain event has probability 1.
Since
PNA one and only one of the events must happen in each trial, the probabilities are normalized,
j P (aj ) = 1. A simplified model of our example trial is the rolling of a die. If the die is fair,
we have the special situation of a uniform probability distribution, i.e., P (aj ) = 1/6 for all j.

Concept 3.1.1 Random variable. A set of random events with their associated probabilities is
called a random variable. If the number of random events is countable, the random variable is
called discrete. In a computer, numbers can be assigned to the events, which makes the random
variable a random number. A series of trials can then be simulated by generating a series of N
pseudo-random numbers that assign the events observed in the N trials. Such simulations are
called Monte Carlo simulations. Pseudo-random numbers obtained from a computer function
need to be adjusted so that they reproduce the given or assumed probabilities of the events.

Problem 3.1 Using the Matlab function rand, which provides uniformly distributed random
numbers in the open interval (0, 1), write a program that simulates throwing a die with six
faces. The outer function should have trial number N as an input and a vector of the numbers of
encountered ones, twos, ... and sixes as an output. It should be based on an inner function that
simulates a single throw of the die. Test the program by determining the difference from the
expectation P (aj ) = 1/6 for ever larger numbers of trials.
18 Probability Theory

3.1.2 Multiple Discrete Random Variables


For two sets of events A and B and their probabilities, we define a joint probability P (aj , bk )
that is the probability of observing both aj and bk in the same trial. An example is the throwing
of two dice, one black and one red, and asking about the probability that the black die shows a 2
and the red die a 3. A slightly more complicated example is the measurement of the individual z
components of spin angular momentum of two coupled spins SA = 5/2 and SB = 5/2. Like
individual probabilities, joint probabilities fall in the closed interval [0, 1]. Joint probabilities are
normalized,
XX
P (a, b) = 1 . (3.2)
a b

Note that we have introduced a brief notation that suppresses indices j and k. This notation is
often encountered because of its convenience in writing.
If we know the probabilities P (a, b) for all NA NB possible combinations of the two events,
we can compute the probability of a single event, for instance a,
X
PA (a) = P (a, b) , (3.3)
b

where PA (a) is the marginal probability of event a.

R The unfortunate term marginal does not imply a small probability. Historically, these
probabilities were calculated in the margins of probability tables [Swe12].

Another quantity of interest is the conditional probability P (a|b) of an event a, provided that
b has happened. For instance, if we call two cards from a full deck, the probability of the second
card being a Queen is conditional on the first card having been a Queen. With the definition for
the conditional probability we have

P (a, b) = P (a|b)PB (b) (3.4)


= P (b|a)PA (a) . (3.5)

Theorem 3.1.2 Bayes theorem. If the marginal probability of event b is not zero, the
conditional probability of event a given b is

P (b|a)PA (a)
P (a|b) = . (3.6)
PB (b)

Bayes theorem is the basis of Bayesian inference, where the probability of proposition a
is sought given prior knowledge (short: the prior) b. Often Bayesian probability is interpreted
subjectively, i.e., different persons, because they have different prior knowledge b, will come to
different assessments for the probability of proposition a. This interpretation is incompatible
with theoretical physics, where, quite successfully, an objective reality is assumed. Bayesian
probability theory can also be applied with an objective interpretation in mind and is nowadays
used, among else, in structural modeling of biomacromolecules to assess agreement of a model
(the proposition) with experimental data (the prior).
In experimental physics, biophysics, and physical chemistry, Bayes theorem can be used
to assign experimentally informed probabilities to different models for reality. For example
assume that a theoretical modeling approach, for instance an MD simulation, has provided a
set of conformations A = {aj } of a protein molecule and associated probabilities PA (aj ). The
probabilities are related, via the Boltzmann distribution, to the free energies of the conformations
3.1 Discrete Probability Theory 19

(this point is discussed later in the lecture course). We further assume that we have a measurement
B with output bk and we know the marginal probability PB (b) of encountering this output for a
random set of conformations of the protein molecule. Then we need only a physical model that
provides the conditional probabilities P (bk |aj ) of measuring bk given the conformations aj and
can compute the probability P (aj |bk ) that the true conformation is aj , given the result of our
measurement, via Bayes theorem. Eq. (3.6). This procedure can be generalized to multiple
measurements. The required P (bk |aj ) depend on measurement errors. The approach allows for
combining possibly conflicting modeling and experimental results to arrive at a best estimate
for the distribution of conformations.
The events associated with two random variables can occur completely independent of each
other. This is the case for throwing two dice: the number shown on the black die does not depend
on the number shown on the red die. Hence, the probability to observe a 2 on the black and a
3 on the red die is (1/6) (1/6) = 1/36. In general, joint probabilities of independent events
factorize into the individual (or marginal) probabilities, which leads to huge simplifications in
computations. In the example of two coupled spins SA = 5/2 and SB = 5/2 the two random
variables mS,A and mS,B may or may not be independent. This is decided by the strength of the
coupling, the preparation of trial T , and the evolution time t before observation.

Concept 3.1.3 Independent variables. If two random variables are independent, the joint
probability of two associated events is the product of the two marginal probabilities,

P (a, b) = PA (a)PB (b) . (3.7)

As a consequence, the conditional probability P (a|b) equals the marginal probability of a (and
vice versa),

P (a|b) = PA (a) . (3.8)

For a set of more than two random variables two degrees of independence can be established,
a weak type of pairwise independence and a strong type of mutual independence. The set
is mutually independent if the marginal probability distribution in any subset, i.e. the set of
marginal probabilities for all event combinations in this subset, is given by the product of the
corresponding marginal distributions for the individual events.1 This corresponds to complete
independence. Weaker pairwise independence implies that the marginal distributions for any pair
of random variables are given by the product of the two corresponding distributions. Note that
even weaker independence can exist within the set, but not throughout the set. Some, but not all
pairs or subsets of random variables can exhibit independence.
Another important concept for multiple random variables is whether or not they are distin-
guishable. In the example above we used a black and a red die to specify our events. If both
dice would be black, the event combinations (a2 , b3 ) and (a3 , b2 ) would be indistinguishable
and the corresponding composite event of observing a 2 and a 3 would have a probability of
1/18, i.e. the product of the probability 1/36 of the basic composite event with its multiplicity 2.
In general, if n random variables are indistinguishable, the multiplicity equals the number of
permutations of the n variables, which is n! = 1 2 (n 1) n.

1As the distributions are vectors and all combinations have to be considered, an outer product must be taken.
20 Probability Theory

3.1.3 Functions of Discrete Random Variables


We consider an event g that depends on two other events a and b. For example, we ask for the
probability that the sum of the numbers shown by the black and red die is g, where g can range
from 2 to 12, given that we know the probabilities P (a, b), which in our example all have the
value 1/36. In general, the probability distribution of random variable G can be computed by
XX
PG (g) = g,G(a,b) P (a, b) , (3.9)
a b

where G(a, b) is an arbitrary function of a and b and the Kronecker delta g,G(a,b) assumes
the value one if g = G(a, b) and zero otherwise. In our example, g = G(a, b) = a + b will
assume the value of 5 for the event combinations (1, 4), (2, 3), (3, 2), (4, 1) and no others. Hence,
PG (5) = 4/36 = 1/9. There is only a single combination for g = 2, hence PG (2) = 1/36, and
there are 6 combinations for g = 7, hence PG (7) = 1/6. Although the probability distributions
for the individual random numbers A and B are uniform, the one for G is not. It peaks at the
value of g = 7 that has the most realizations. Such peaking of probability distributions that
depend on multiple random variables occurs very frequently in statistical mechanics. The peaks
tend to become the sharper the larger the number of random variables that contribute to the sum.
If this number N tends to infinity, the distribution of the sum g is so sharp that the distribution
width (to be specified below) is smaller than the error in the measurement of the mean value
g/N (see Section 3.1.5). This effect is the very essence of statistical thermodynamics: Although
quantities for a single molecule may be broadly distributed and unpredictable, the mean value for
a large number of molecules, lets say 1018 of them, is very well defined and perfectly predictable.
In a numerical computer program, Eq. (3.9) for only two random variables can be implemented
very easily by a loop over all possible values of g with inner loops over all possible values of a
and b. Inside the innermost loop, G(a, b) is computed and compared to loop index g to add or
not add P (a, b) to the bin corresponding to value g. Note however that such an approach does
not carry to large numbers of random variables, as the number of nested lops increases with
the number of random variables and computation time thus increases exponentially. Analytical
computations are simplified by the fact that g,G(a,b) usually deviates from zero only within
certain ranges of the summation indexes j (for a) and k (for b). The trick is then to find the proper
combinations of index ranges.
Problem 3.2 Compute the probability distribution for the sum g of the numbers shown by two
dice in two ways. First, write a computer program using the approach sketched above. Second,
compute the probability distribution analytically by making use of the uniform distribution for
the individual events (P (a, b) = 1/36 for all a, b. For this, consider index ranges that lead to a
given value of the sum g.2

3.1.4 Discrete Probability Distributions


In most cases random variables are compared by considering the mean values and widths of
their probability distributions. As a measure of the width, the standard deviation of the values
from the mean value is used, which is the square root of the variance 2 . The concept can be
generalized by considering functions f (A) of the random variable. In the following expressions,
f (A) = A provides the mean value and standard deviation of the original random variable A.

Concept 3.1.4 Mean value and standard deviation. For any function F (A) of a random

2The solution of the second task can be found in [Swe12].


3.1 Discrete Probability Theory 21

variable A, the mean value hF i is given by,


X
hF i = F (a)PA (a) . (3.10)
a

The standard deviation, which characterizes the width of the distribution of the function values
f (a), is given by,
s
X
= (F (a) hF i)2 PA (a) . (3.11)
a

The mean value is the first moment of the distribution, with the nth moment being defined by
X
hF n i = F n (a)PA (a) . (3.12)
a

The nth central moment is


X
h(F hF i)n i = (F (a) hF i)n PA (a) . (3.13)
a

For the variance, which is the second central moment, we have

2 = hF 2 i hF i2 . (3.14)

Assume that we know the mean values for functions F (A) and G(B) of two random variables
as well as the mean value hF Gi of their product, which we can compute if the joint probability
function P (a, b) is known. We can then compute a correlation function

RF G = hF Gi hF ihGi , (3.15)

which takes the value of zero, if F and G are independent random numbers.
Problem 3.3 Compute the probability distribution for the normalized sum g/M of the numbers
obtained on throwing M dice in a single trial. Start with M = 1 and proceed via M =
10, 100, 1000 to M = 10000. Find out how many Monte Carlo trials N you need to guess the
converged distribution. What is the mean value hg/M i? What is the standard deviation g ? How
do they depend on N ?

3.1.5 Probability Distribution of a Sum of Random Numbers


If we associate the random numbers with N molecules, identical or otherwise, we will often
need to compute the sum over all molecules. This generates a new random number

N
X
S= Fj , (3.16)
j=1

whose mean value is the sum of the individual mean values,

N
X
hSi = hFj i . (3.17)
j=1
22 Probability Theory

If motion of the individual molecules is uncorrelated, the individual random numbers Fj are
independent. It can then be shown that the variances add [Swe12],
N
X
S2 = j2 (3.18)
j=1

For identical molecules, all random numbers have the same mean hF i and variance F2 and
we find

hSi = N hF i (3.19)
S2 = N F2 (3.20)

S = N F . (3.21)

This result relates to the concept of peaking of probability distributions for a large number of
molecules that was introduced above on the example of the probability distribution for sum of
the numbers shown by two dice. The width of the distribution normalized to its mean value,
S 1 F
= , (3.22)
hSi N hF i

scales with the inverse square root of N . For 1018 molecules, this relative width of the distribution
is one billion times smaller than for a single molecule. Assume that for a certain physical quantity
of a single molecule the standard deviation is as large as the mean value. No useful prediction
can be made. For a macroscopic sample, the same quantity can be predicted with an accuracy
better than the precision that can be expected in a measurement.

3.1.6 Binomial Distribution


We consider the measurement of the z component of spin angular momentum for an ensemble of
N spins S = 1/2.3 The random number associated with an individual spin can take only two
values, ~/2 or +~/2. Additive and multiplicative constants can be taken care of separately
and we can thus represent each spin by a random number A that assumes the value a = 1 (for
mS = +1/2) with probability P and, accordingly, the value a = 0 (for mS = 1/2) with
probability 1 P . This is a very general problem, which also relates to the second postulate of
Penrose (see Section 2.3.1). A simplified version with P = 1 P = 0.5 is given by N flips of a
fair coin. A fair coin or a biased coin with P 6= 0.5 can be easily implemented in a computer,
for instance by using a = floor(rand+P) in Matlab . For the individual random numbers we
find hAi = P and A 2 = P (1 P ), so that the relative standard deviation for the ensemble with
p
N members becomes S /hSi = (1 P )/(N P ).4
To compute the explicit probability distribution of the sum of the random numbers for the
whole ensemble, we realize that the probability of a subset of n ensemble members providing a 1
and N n ensemble members providing a 0 is P n (1 P )N n . The value of the sum associated
with this probability is n.
Now we still need to consider the phenomenon already encountered for the sum of the
numbers on the black and red dice: Different numbers n have different multiplicities. We
have N ! permutations of the ensemble members. Let us assign a 1 to the first n members of
each permutation. For our problem, it does not matter in which sequence these n members are
numbered and it does not matter in which sequence the remaining N n members are numbered.
3Here, each individual spin is considered as a system and the N spins as an ensemble of identical systems
4Thanks to Takuya Segawa for pointing out a mistake in this expression in [Swe12]
3.1 Discrete Probability Theory 23

Hence, we need to divide the total number of permutations N ! by the numbers of permutations
in each subset, n! and (N n)! for the first and second subset, respectively. The multiplicity that
we need is the number of combinations of N elements to the nth class, which is thus given by
the binomial coefficient,
 
N N!
= , (3.23)
n n!(N n)!
providing the probability distribution
 
N
PS (n) = P n (1 P )N n . (3.24)
n

10-3
0.03 4

0.025
A B

[G(n) - P(n)]/max(P(n))
2
0.02
P(n)

0.015 0

0.01
-2
0.005

0 -4
0 200 400 600 800 1000 0 200 400 600 800 1000
n n

Figure 3.1: Gaussian approximation of the binomial distribution. (A) Gaussian approximation (red
dashed line) and binomial distribution (black solid line) for P = 0.37 and N = 1000. (B) Error of the
Gaussian approximation relative to the maximum value of the binomial distribution.

For large values of N the binomial distribution tends to a Gaussian distribution,


(s hsi)2
 
1
G(s) = exp . (3.25)
2 2 2 2
As we already know the mean value hsi = hni = N P and variance S2 = N P (1 P ), we can
immediately write down the approximation
(n P N )2
 
1
PS (n) p exp = G(n) . (3.26)
2P (1 P )N 2P (1 P )N
As shown in Figure 3.1 the Gaussian approximation of the binomial distribution is quite good
already at N = 1000.
In fact, the Gaussian (or normal) distribution is a general distribution for the arithmetic mean
of a large number of independent random variables:

Concept 3.1.5 Central limit theorem. Suppose that a large number N of observations has
been made with each observation corresponding to a random number that is independent
from the random numbers of the other observations. According to the central limit theorem,
the mean value hSi/N of the sum of all these random numbers is approximately normally
distributed, regardless of the probability distribution of the individual random numbers, as
long all the probability distributions of all individual random numbers are identical.a The
central limit theorem applies, if each individual random variable has a well-defined mean value
24 Probability Theory

(expectation value) and a well-defined variance. These conditions are fulfilled for statistically
regular trials T .
a
If the individual random numbers are not identically distributed, the theorem will still apply, if Lyapunovs
condition or Lindebergs condition is fulfilled. See the very useful and detailed Wikipedia article on the Central
limit theorem for more information and proofs.

3.1.7 Stirlings Formula


The number N ! of permutations increases very fast with N , leading to numerical overflow
in calculators and computers at values of N that correspond to nanoclusters rather than to
macroscopic samples. Even binomial coefficients, which grow less strongly with increasing
ensemble size, cannot be computed with reasonable precision for N  1000. Furthermore, the
factorial N ! is difficult to handle in calculus. The scaling problem can be solved by taking the
logarithm of the factorial,
N N
!
Y X
ln N ! = ln n = ln n . (3.27)
n=1 n=1

Concept 3.1.6 Stirlings formula.For large numbers N the natural logarithm of the factorial
can be approximated by Stirlings formula

ln N ! N ln N N + 1 , (3.28)

which amounts to the approximation

N ! N N exp(1 N ) (3.29)

for the factorial itself. For large numbers N it is further possible to neglect 1 in the sum and
approximate ln N ! N ln N N .

The absolute error of this approximation for N ! looks gross and increases fast with increasing
N , but because N ! grows much faster, the relative error becomes insignificant already at moderate
N . For ln N ! it is closely approximated by 0.55/N . In fact, an even better approximation has
been found by Gosper [Swe12],
  
1 1
ln N ! N ln N N + ln 2N + . (3.30)
2 3
Gospers approximation is useful for considering moderately sized systems, but note that several
of our other assumptions and approximations become questionable for such systems and much
care needs to be taken in interpreting results. For the macroscopic systems, in which we are
mainly interested here, Stirlings formula is often sufficiently precise and Gospers is not needed.
Slightly better than Stirlings original formula, but still a simple approximation is
 N
N
N ! 2N . (3.31)
e

3.2 Continuous Probability Theory


3.2.1 Probability Density
Although the outcomes of measurements can be discretized, and in fact, are invariably discretized
when storing the data, in theory it is convenient to work with continuous variables where physical
3.2 Continuous Probability Theory 25

quantities are assumed to be continuous. For instance, spatial coordinates in phase space are
assumed to be continuous, as are the momentum coordinates for translational motion in free
space.
To work with continuous variables, we assume that an event can return a real number instead
of an integer index. The real number with its associated probability density is a continuous
random number. Note the change from assigning a probability to an event to assigning a
probability density. This is necessary as real numbers are not countable and thus the number of
possible events is infinite. If we want to infer a probability in the usual sense, we need to specify
an interval [l, u] between a lower bound l and an upper bound u. The probability that trial T will
turn up a real number in this closed interval is given by
Z u
P ([l, u]) = (x)dx . (3.32)
l

The probability density must be normalized,


Z
(x)dx = 1 . (3.33)

A probability density distribution can be characterized by its moments.

Concept 3.2.1 Moment analysis. The nth moment of a probability density distribution is
defined as,
Z
n
hx i = xn (x)dx . (3.34)

The first moment is the mean of the distribution. With the mean hxi, the central moments are
defined
Z
n
h(x hxi) i = (x hxi)n (x)dx . (3.35)

The second central moment is the variance x2 and its square root x is the standard deviation.

Probability density is defined along some dimension x, corresponding to some physical


quantity. The average of a function F (x) of this quantity is given by
Z
hF (x)i = F (x)(x)dx . (3.36)

R In many books and articles, the same symbol P is used for probabilities and probability
densities. This is pointed out by Swendsen [Swe12] who decided to do the same, pointing
out that the reader must learn to deal with this. In the next section he goes on to confuse
marginal and conditional probability densities with probabilities himself. In these lecture
notes we use P for probabilities, which are always unitless, finite numbers in the interval
[0, 1] and for probability densities, which are always infinitesimally small and may have
a unit. Students are advised to keep the two concepts apart, which means using different
symbols.

Computer representations of probability densities by a vector or array are discretized. Hence,


the individual values are finite. We now consider the problem of generating a stream of
random numbers that conforms to a given discretized probability density ~ [Pre+97]. Modern
programming languages or mathematical libraries include functions that provide uniformly
26 Probability Theory

10 -3
12
1
A B

Probability density [a.u.]


10
Cumulative probability

0.8 0.75
8
0.45
0.6
6

0.4 4

0.2 2

0 0

-3 -2 -1 0 1 2 3 -3 -2 -1 0 1 2 3
x x

Figure 3.2: Generation of random numbersRthat conform to a given probability density distribution. (A)
x
Cumulative probability distribution P (x) = ()d for (x) = exp(x4 ) (blue). A pseudo-random
number with uniform distribution in (0, 1), here 0.75, selects the ordinate of P (x) (red dashed horizontal
line). The corresponding abscissa, here x = 0.45 (red dashed vertical line), is an instance of a random
number with probability density distribution (x). (B) Distribution of 105 random numbers (grey line)
and target probability density distribution (x) = exp(x4 ) (black line).

distributed pseudo-random numbers in the interval (0, 1) (Matlab : rand) or pseudo-random


numbers with a Gaussian (normal) distribution with mean 0 and standard deviation 1 (Matlab :
randn). A stream of uniformly distributed pseudo-random numbers in (0, 1) can be transformed
to a stream of numbers with probability density conforming to ~ by selecting for each input
number the abscissa where the cumulative sum of ~ (Matlab : cumsum(rho)) most closely
matches the input number (Figure 3.2). Note that ~ must be normalized (Matlab : rho =
rho/sum(rho)). Since a random number generator is usually called very often in a Monte Carlo
simulation, the cumulative sum cumsum_rho should be computed once for all before the loop
over all trials. With this, generation of the abscissa index poi becomes a one-liner in Matlab :
[~,poi] = min(abs(cumsum_rho - rand));5
Coming back to physical theory, the concept of probability density can be extended to
multiple dimensions, for instance to the 2F = 2f M dimensions of phase space. Probability
then becomes a volume integral in this hyperspace. A simple example of a multidimensional
continuous problem is the probability of finding a classical particle in a box. The probability
to find it at a given point is infinitely small, as there are infinitely many of such points. The
probability density is uniform, since all points are equally likely for a classical (unlike a quantum)
particle. With the volume V of the box, this uniform probability density is 1/V ifR we have a
single particle in the box. This follows from the normalization condition, which is dV = 1.
Note that a probability density has a unit, in our example m3 . In general, the unit is the inverse
of the product of the units of all dimensions.
The marginal probability density for a subset of the events is obtained by integrating out the
other events. Let us assume a particle in a two-dimensional box with dimensions x and y and ask
about the probability density along x. It is given by
Z
x (x) = (x, y)dy . (3.37)

5This one-liner may cause efficiency problems if computational effort per trial besides random number generation
is small.
3.2 Continuous Probability Theory 27

Likewise, the conditional probability density (y|x) is defined at all points where x (x) 6= 0,

(x, y)
(y|x) = . (3.38)
x (x)

If two continuous random numbers are independent, their joint probability density is the product
of the two individual probability densities,

(x, y) = x (x)y (y) . (3.39)

A B

Figure 3.3: Monte Carlo simulation of a two-dimensional probability density distribution. (A) Two-
dimensional probability density distribution corresponding to the first-order membrane function used
in the Matlab logo. (B) Distribution of 107 random numbers conforming to the probability density
distribution shown in (A).

Problem 3.4 Write a Matlab program that generates random numbers conforming to a two-
dimensional probability density distribution mem that resembles the Matlab logo (see Figure
3.3). The (not yet normalized) distribution mem is obtained with the function call L =
membrane(1,resolution,9,9);. Hint: You can use the reshape function to generate a
vector from a two-dimensional array as well as for reshaping a vector into a two-dimensional
array. That way the two-dimensional problem (or, in general, a multi-dimensional problem) can
be reduced to the problem of a one-dimensional probability density distribution.

3.2.2 Selective Integration of Probability Densities


We already know how to compute probability from probability density for a simply connected
parameter range. Such a range can be an interval [l, u] for a probability density depending on
only one parameter x or a simply connected volume element for a probability density depending
on multiple parameters. In a general problem, the points that contribute to the probability of
interest may not be simply connected. If we can find a function g(x) that is zero at the points
that should contribute, we can solve this problem with the Dirac delta function, which is the
continuous equivalent of the Kronecker delta that was introduced above.

Concept 3.2.2 Dirac delta function. The Dirac delta function is a generalized function with
the following properties
1. RThe function (x) is zero everywhere except at x = 0.

2. (x)dx = 1.
The function can be used to select the value f (x0 ) of another continuous function f (x),
Z
f (x0 ) = f (x)(x x0 )dx . (3.40)

28 Probability Theory

This concept can be used, for example, to compute the probability density of a new random
variable s that is a function of two given random variables x and y with given joint probability
density (x, y). The probability density (s) corresponding to s = f (x, y) is given by
Z Z
(s) = (x, y) (s f (x, y)) dxdy . (3.41)

Note that the probability density (s) computed that way is automatically normalized.

3.2.3 Sum of Two Continuous Random Numbers

y
a 8
c
1
r(x)

6
0 6 x
4
b 1/36
1
r(y)

0 6 y 0 4 6 8 x

Figure 3.4: Probability density distributions for two continuous random numbers x and y that are
uniformly distributed in the interval [0, 6] and have zero probability density outside this interval. a)
Marginal probability density x (x). b) Marginal probability density y (y). c) Joint probability density
(x, y). In the light blue area, = 1/36, outside = 0. The orange line corresponds to s = 4 and the
green line to s = 8.

We now use the concept of selective integration to compute the probability density (s) for
the sum s = x + y of the numbers shown by two continuous dice, with each of them having a
uniform probability density in the interval [0, 6] (see Fig. 3.4). We have
Z Z
(s) = (x, y) (s (x + y)) dydx (3.42)

Z 6Z 6
1
= (s (x + y)) dydx . (3.43)
36 0 0

The argument of the delta function in the inner integral over y can be zero only for
0 s x 6, since otherwise no value of y exists that leads to s = x + y. It follows that x s
and x s 6. For s = 4 (orange line in Fig. 3.4c) the former condition sets the upper limit of
the integration. Obviously, this is true for any s with 0 s 6. For s = 8 (orange line in Fig.
3.4c) the condition x s 6 sets the lower limit of the integration, as is also true for any s with
6 s 12. The lower limit is 0 for 0 s 6 and the upper limit is 6 for 6 s 12. Hence,
Z s
1 s
(s) = ds = for s 6 , (3.44)
36 0 36
3.2 Continuous Probability Theory 29

and
6
12 s
Z
1
(s) = ds = for s 6 . (3.45)
36 s6 36

From the graphical representation in Fig. 3.4c it is clear that (s) is zero at s = 0 and s = 12,
assumes a maximum of 1/6 at s = 6, increases linearly between s = 0 and s = 6 and decreases
linearly between s = 6 and s = 12.
Statistical Ensembles
Concept of an Ensemble
Ergodicity
Microcanonical Ensemble
Canonical Ensemble
Boltzmann Distribution
Equipartition Theorem
Internal Energy and Heat Capacity of the Canoni-
cal Ensemble
Grand Canonical Ensemble

4 Classical Ensembles

4.1 Statistical Ensembles


4.1.1 Concept of an Ensemble
Probability densities in phase space cannot be computed by considering only a single system
at a single instant in time. Such a system will be in some random microstate, but what we
need is the statistics of such microstates. This problem was solved by Gibbs, who considered
ensembles that consist of a very large number of identical systems in possibly different microstates.
The microstates for a system with M molecules with f degrees of freedom each are points in
2f M -dimensional phase space. If we have information on the probability density assigned to
such points, we can use probability theory to compute thermodynamical state functions.

4.1.2 Ergodicity
Instead of considering a large ensemble of systems at the same time (ensemble average), we
could also consider a long trajectory of a single system in phase space. The single system
will go through different microstates and if we observe it for a sufficiently long time, we might
expect that it visits all accessible points in phase space with a frequency that corresponds to the
associated probability density. This idea is the basis of analyzing MD trajectories in terms of
thermodynamic state functions. The ensemble average hAi is replaced by the time average A.
We assume

hAi = A . (4.1)

Systems where this assumption holds are called ergodic systems.


Often, experiments are performed on a large ensemble of identical systems. An example
is a spectroscopic experiment on a dilute solution of chromophores: Each chromophore can
be considered as an individual system and their number may be of the order of 1010 or higher.
In some cases an equivalent experiment can be performed on a single chromophore, but such
single-molecule experiments require many repetitions and measure a time-average. The results
of ensemble and single-molecule experiments are equivalent if the system is ergodic and the
measurement time in the single-molecule experiment is sufficiently long.
Whether or not a system is ergodic depends on kinetic accessibility of the whole thermody-
namically accessible phase space. We shall see later that thermodynamic accessibility is related
32 Classical Ensembles

to temperature and to the energy assigned to points in phase space. Points are accessible if
their energy is not too much higher than the energy minimum in phase space. Whether a single
dynamic system visits all these points at the same given temperature- and what time it needs
to sample phase space- depends on energy barriers. In MD simulations sampling problems are
often encountered, where molecular conformations that are thermodynamically accessible are
not accessed within reasonable simulation times. A multitude of techniques exists for alleviating
such sampling problems, none of them perfect. In general, time-average methods, be they
computational or experimental, should be interpreted only with care in terms of thermodynamics.
In this lecture course we focus on ensemble-average methods, which suffer from a loss in dynamic
information, but get the thermodynamic state functions right.

4.2 Microcanonical Ensemble


Assume that we have an isolated system with N particles in a fixed volume V . Because the
system is isolated, the total energy E must also be fixed. If we know that the energy must be in
an interval [E, E + E] the probability density in phase space must be zero everywhere outside
the region between the two hypersurfaces with constant energies E and E + E. We call this
region the energy shell in which the system is confined. If the system is in equilibrium, i.e., the
probability density is stationary, must be uniform in this energy shell, i.e., it must not depend
on p and q within this shell. We can see this from the Liouville equation (2.8), whose left-hand
side must be zero for a stationary probability density. The Poisson bracket on the right-hand side
will vanish if is uniform.1

Concept 4.2.1 Microcanonical ensemble. An ensemble with a constant number N of particles


in a constant volume V and with constant total energy E has a uniform probability density
mc in the part of phase space, where it can reside, which is the energy hypersurface at energy
E. Such an ensemble is called a microcanonical ensemble.
We are left with computing this constant probability density mc . As the energy is given by
the Hamiltonian function H(p, q), we can formally write mc for an infinitely thin energy shell
(E 0) as
1
mc = (E H(p, q)) , (4.2)
(E)
where the statistical weight depends on energy, volume, and number of particles N , but at
constant energy does not depend on momentum p or spatial coordinates q. Since the probability
density is normalized, we have
Z Z
(E) = (E H(p, q)) dqdp . (4.3)

The probability density in phase space of the microcanonical ensemble is thus relatively
easy to compute. However, the restriction to constant energy, i.e. to an isolated system,
severely limits application of the microcanonical ensemble. To see this, we consider the simplest
system, an electron spin S = 1/2 in an external magnetic field B0 . This system is neither
classical nor describable in phase space, but it will nicely serve our purpose. The system has
a state space consisting of only two states |i and |i with energies  = ~ge B B0 /2 and
 = ~ge B B0 /2.2 In magnetic resonance spectroscopy, one would talk of an ensemble of
1It is more tricky to argue that it will only vanish if is uniform. However, as the individual particles follow
random phase space trajectories, it is hard to imagine that the right-hand side could be stationary zero unless is
uniform.
2Where ge is the g value of the free electron and B the Bohr magneton.
4.3 Canonical Ensemble 33

isolated spins, if the individual spins do not interact with each other. We shall see shortly
that this ensemble is not isolated in a thermodynamical sense, and hence not a microcanonical
ensemble.
The essence of the microcanonical ensemble is that all systems in the ensemble have the same
energy E, this restricts probability density to the hypersurface with constant E. If our ensemble
of N spins would be a microcanonical ensemble, this energy would be either E = ~ge B B0 /2
or E = ~ge B B0 /2 and all spins in the ensemble would have to be in the same state, i.e., the
ensemble would be in a pure state. In almost any experiment on spins S = 1/2 the ensemble
is in a mixed state and the populations of states |i and |i are of interest. The system is not
isolated, but, via spin relaxation processes, in thermal contact with its environment. To describe
this situation, we need another type of ensemble.

4.3 Canonical Ensemble


Equilibrium thermodynamics describes systems that are in thermal equilibrium. In an ensemble
picture, this can be considered by assuming that the system is in contact with a very large for
mathematical purposes infinitely large heat bath. Because of this, the individual systems in the
ensemble can differ in energy. However, the probability density distribution in phase space or
state space must be consistent with constant temperature T , which is the temperature of the heat
bath. In experiments, it is the temperature of the environment.

Concept 4.3.1 Canonical ensemble. An ensemble with a constant number N of particles in


a constant volume V and at thermal equilibrium with a heat bath at constant temperature T
can be considered as an ensemble of microcanonical subensembles with different energies i .
The energy dependence of probability density conforms to the Boltzmann distribution. Such
an ensemble is called a canonical ensemble.

R Because each system can exchange heat with the bath and thus change its energy, systems
will transfer between subensembles during evolution. This does not invalidate the idea of
microcanonical subensembles with constant particle numbers Ni . For a sufficiently large
ensemble at thermal equilibrium the ni are constants of motion.

There are different ways of deriving the Boltzmann distribution. Most of them are rather
abstract and rely on a large mathematical apparatus. The derivation gets lengthy if one wants
to create the illusion that we know why the constant introduced below always equals 1/kB T ,
where kB = R/NAv is the Boltzmann constant, which in turn is the ratio of the universal gas
constant R and the Avogadro constant NAv . Here we follow a derivation [WF12] that is physically
transparent and relies on a minimum of mathematical apparatus that we have already introduced.

4.3.1 Boltzmann Distribution


Here we digress from the ensemble picture and use a system of N particles that may exist in
r different states with energies i with i = 0 . . . r. The number of particles with energy i is
Ni . The particles do not interact, they are completely independent from each other. We could
therefore associate theses particles with microcanonical subensembles of a canonical ensemble,
but the situation is easier to picture with particles. The probability Pi = Ni /N to find a particle
with energy i can be associated with the probability density for the microcanonical subensemble
at energy i . The difference between this simple derivation and the more elaborate derivation for a
canonical ensemble is thus essentially the difference between discrete and continuous probability
theory. We further assume that the particles are classical particles and thus distinguishable.
34 Classical Ensembles

To compute the probability distribution P (i) = Ni /N , we note that


r1
X
Ni = N (4.4)
0

and
r1
X
Ni i = E , (4.5)
0

where E is a constant total energy of the system. We need to be careful in interpreting the
latter equation in the ensemble picture. The quantity E corresponds to the energy of the whole
canonical ensemble, which is indeed a constant of motion, if we consider a sufficiently large
number of systems in contact with a thermal bath. We can thus use our simple model of N
particles for guessing the probability density distribution in the canonical ensemble.
What we are looking for is the most likely distribution of the N particles on the r energy
levels. This is equivalent to putting N distinguishable balls into r boxes. We did already solve
the problem of distributing N objects to 2 states when considering the binomial distribution in
 of a configuration with n objects in the first state and N n
Section 3.1.6. The statistical weight
objects in the second state was N n . With this information we would already be able to solve the
problem of a canonical ensemble of N spins S = 1/2 in thermal contact with the environment,
disregarding for the moment differences between classical and quantum statistics (see Section
6.2).
Coming back to N particles and r energy levels, we still have N ! permutations. If we assign
the first N0 particles to the state with energy 0 , the next N1 particles to 1 and so on, we need
to divide each time by the number of permutations Ni ! in the same energy state, because the
sequence of particles with the same energy does not matter. We call the vector of the occupation
numbers Ni a configuration. The configuration specifies one particular macrostate of the system
and the relative probability of the macrostates is given by their statistical weights,
N!
= . (4.6)
N0 !N1 ! . . . Nr1 !
The most probable macrostate is the one with maximum statistical weight . Because of the
peaking of probability distributions for large N , we need to compute only this most probable
macrostate; it is representative for the whole ensemble. Instead of maximizing we can as well
maximize ln , as the natural logarithm is a strictly monotonous function. This allows us to
apply Stirlings formula,
r1
X
ln = ln N ! ln Ni ! (4.7)
0
r1
X r1
X
= N ln N N + 1 Ni ln Ni + Ni r . (4.8)
0 0

By inserting Eq. (4.4) we find


r1
X
ln = N ln N Ni ln Ni + 1 r . (4.9)
0

Note that the second term on the right-hand side of Eq. (4.9) has some similarity to the entropy
of mixing, which suggests that ln is related to entropy.
4.3 Canonical Ensemble 35

At the maximum of ln the derivative of ln with respect to the Ni must vanish,


X X X X
0= Ni ln Ni = (Ni ln Ni + Ni ln Ni ) = Ni + ln Ni Ni . (4.10)
i i i i

In addition, we need to consider the boundary conditions of constant particle number, Eq. (4.4),
X
N = Ni = 0 (4.11)
i

and constant total energy, Eq. (4.5),


X
E = i Ni = 0 . (4.12)
i

It might appear that Eq. (4.11) could be used to cancel a term in Eq. (4.10), but this would
be wrong as Eq. (4.11) is a constraint that must be fulfilled separately. For the constrained
maximization we can use the method of Lagrange multipliers.

Concept 4.3.2 Lagrange multipliers. The maximum or minimum of a function f (x1 . . . , xn )


of n variables is a stationary point that is attained at
n  
X f
f = xi = 0 . (4.13)
xi xk 6=xi
i=1

We now consider the case where the possible sets of the n variables are constrained by c
additional equations

gj (x1 , x2 , . . . , xn ) = 0 , (4.14)

where index j runs over the c constraints (j = 1 . . . c). Each constraint introduces another
equation of the same form as the one of Eq. (4.13),
n  
X gj
gj = xi = 0 . (4.15)
xi xk 6=xi
i=1

The constraints can be introduced by multiplying each of the c equations by a multiplier j


and subtracting it from the equation for the stationary point without the constraints,

n   c  
X
f
X gj
L= j xi . (4.16)
xi xk 6=xi xi xk 6=xi
i=1 j=1

If a set of variables {x0,1 . . . , x0,n } solves the constrained problem then there exists a set
{0,1 . . . 0,r } for which {x0,1 , x0,2 , . . . , x0,n } also corresponds to a stationary point of the
Lagrangian function L(x1 , . . . , xn , 1 , . . . r ). Note that not all stationary points of the
Lagrangian function are necessarily solutions of the constrained problem. This needs to be
checked separately.

With this method, we can write


X X X X
0= Ni + ln Ni Ni + Ni + i Ni (4.17)
i i i i
X
= Ni (1 + ln Ni + + i ) . (4.18)
i
36 Classical Ensembles

The two boundary conditions fix only two of the population numbers Ni . We can choose the
multipliers and in a way that (1 + ln Ni + + i ) = 0 for these two Ni , which ensures that
the partial derivatives of ln with respect to these two Ni vanishes. The other r 2 population
numbers can, in principle, be chosen freely, but again we must have

1 + ln Ni + + i = 0 (4.19)

for all i to make sure that we find a maximum with respect to variation of any of the r population
numbers. This gives

Ni = ei (4.20)

with = e(1+) . We can eliminate by using Eq. (4.4),


X X
Ni = ei = N , (4.21)
i i

giving
N
=P i
, (4.22)
ie

and finally leading to

Ni ei
P (i) = = P  . (4.23)
N ie
i

For many problems in statistical thermodynamics, the Lagrange multiplier is related to


the chemical potential by = /(kB T ). The Lagrange multiplier must have the reciprocal
dimension of an energy, as the exponent must be dimensionless. As indicated above, we cannot
at this stage prove that is the same energy for all problems of the type that we have posed here,
let alone for all of the analogous problems of canonical ensembles. The whole formalism can
be connected to phenomenological thermodynamics via Maxwells kinetic gas theory (see also
Section 4.3.2). For this problem one finds
1
= . (4.24)
kB T

Concept 4.3.3 Boltzmann distribution. For a classical canonical ensemble with energy levels
i the probability distribution for the level populations is given by the Boltzmann distribution

Ni ei /kB T
Pi = = P  /k T . (4.25)
N ie
i B

The sum over states


X
Z(N, V, T ) = ei /kB T (4.26)
i

required for normalization is called canonical partition function.a The partition function is a
thermodynamical state function.
a
The dependence on N and V arises, because these parameters influence the energy levels

For the partition function, we use the symbol Z relating to the German term Zustandssumme,
which is a more lucid description of this quantity.
4.3 Canonical Ensemble 37

4.3.2 Equipartition Theorem


Comparison of Maxwells kinetic theory of gases with the state equation of the ideal gas
from phenomenological thermodynamics provides a mean kinetic energy of a point particle of
hkin i = 3kB T /2. This energy corresponds to

1 1 2
trans = mv 2 = p , (4.27)
2 2m
i.e., it is quadratic in the velocity coordinates of dynamic space or the momentum coordinates of
phase space. Translational energy is distributed via three degrees of freedom, as the velocities or
momenta have components along three pairwise orthogonal directions in space. Each quadratic
degree of freedom thus contributes a mean energy of kB T /2.
If we accept that the Lagrange multiplier assumes a value 1/kB T , we find a mean energy
kB T of an harmonic oscillator in the high-temperature limit [WF12]. Such an oscillator has two
degrees of freedom that contribute quadratically in the degrees of freedom to energy,

1 1
vib = v 2 + f x2 , (4.28)
2 2
where is the reduced mass and f the force constant. The first term contributes to kinetic energy,
the second to potential energy. In the time average, each term contributes the same energy and
assuming ergodicity this means that each of the two degrees of freedom contributes with kB T /2
to the average energy of a system at thermal equilibrium.
The same exercise can be performed for rotational degrees of freedom with energy

1
rot = I 2 , (4.29)
2
where I is angular momentum and angular frequency. Each rotational degree of freedom,
being quadratic in again contributes a mean energy of kB T /2.
Based on Eq. (4.23) it can be shown [WF12] that for an energy

f
X
i = 0 + 1 + 2 + . . . = k , (4.30)
k=1

where index k runs over the individual degrees of freedom, the number of molecules that
contribute energy k does not depend on the terms j with j 6= k. It can be further shown that

1
hk i = (4.31)
2

for all terms that contribute quadratically to energy.3


This result has two consequences. First, we can generalize = 1/kB T , which we strictly
knew only for translational degrees of freedom, to any canonical ensemble for which all individual
energy contributions are quadratic along one dimension in phase space. Second, we can formulate
the
Theorem 4.3.4 Equipartition theorem. Each degree of freedom, whose energy scales quadrat-
ically with one of the coordinates of state space, contributes a mean energy of kB T /2.

3The condition of a quadratic contribution arises from an assumption that is made when integrating over the
corresponding coordinate.
38 Classical Ensembles

R The equipartition theorem applies to all degrees of freedom that are activated. Translational
degrees of freedom are always activated and rotational degrees of freedom are activated
at ambient temperature, which corresponds to the high-temperature limit of rotational
dynamics. To vibrational degrees of freedom the equipartition theorem applies only in the
high-temperature limit. In general, the equipartition theorem fails for quantized degrees of
freedom if the quantum energy spacing is comparable to kB T /2 or exceeds this value. We
shall come back to this point when discussing the vibrational partition function.

4.3.3 Internal Energy and Heat Capacity of the Canonical Ensemble


The internal energy u of a system consisting of N particles that are distributed to r energy levels
can be identified as the total energy E of the system considered in Section (4.3.1). Using Eqs.
(4.5) and (4.25) we find
i /kB T i ei /kB T
P P
i i e
u = N P  /k T = N i . (4.32)
ie
i B Z
The sum in the numerator can be expressed by the partition function, since
dZ 1 X i /kB T
= i e . (4.33)
dT kB T 2
i

Thus we obtain
1 dZ d ln Z
u = N kB T 2 = N kB T 2 . (4.34)
Z dT dT
Again the analogy of our simple system to the canonical ensemble holds. At this point we
have computed one of the state functions of phenomenological thermodynamics from the set of
energy levels. The derivation of the Boltzmann distribution has also indicated that ln , and thus
the partition function Z are probably related to entropy. We shall see in Section 5.2 that this is
indeed the case and that we can compute all thermodynamic state functions from Z.
Here we can still derive the heat capacity cV at constant volume, which is the partial derivative
of internal energy with respect to temperature. To that end we note that the partition function for
the canonical ensemble relates to constant volume and constant number of particles.
     
u 2 ln Z ln Z
cV = =N kB T = N kB (4.35)
T V T T V T 1/T V
   
[ ln Z/1/T ] N kB [ ln Z/1/T ]
= N kB = (4.36)
T V T2 1/T V
 2 
kB ln z
= 2 . (4.37)
T (1/T )2 V
In the last line of Eq (4.37) we have substituted the molecular partition function Z by the partition
function for the whole system, ln z = N ln Z. Note that this implies a generalization. Before, we
were considering a system of N identical particles. Now we implicitly assume that Eq. (4.37),
as well as u = kB T 2 ddT ln z
will hold for any system, as long as we correctly derive the system
partition function z.
We note here that the canonical ensemble describes a closed system that can exchange heat
with its environment, but by definition it cannot exchange work, because its volume V is constant.
This does not present a problem, since the state functions can be computed at different V . In
particular, pressure p can be computed from the partition function as well (see Section 5.2).
However, because the canonical ensemble is closed, it cannot easily be applied to all problems
that involve chemical reactions. For this we need to remove the restriction of a constant number
of particles in the systems that make up the ensemble.
4.4 Grand Canonical Ensemble 39

4.4 Grand Canonical Ensemble


For the description of an open system in the thermodynamical sense, i.e., a system that can
exchange not only heat, but also matter with its environment, we need to replace particle number
N with another constant of motion. If we would fail to introduce a new constant of motion, we
would end up with a system that is not at equilibrium and thus cannot be fully described by
time-independent state functions. If we assume that the system is in chemical as well as thermal
equilibrium with its environment, the new constant of motion is the chemical potential , or more
precisely, a vector of the chemical potentials k of all components.

Concept 4.4.1 Grand canonical ensemble. An ensemble with constant chemical potential
k of all components, and constant volume V that is at thermal equilibrium with a heat
bath at constant temperature T and in chemical equilibrium with its environment is called
a grand canonical ensemble. It can be considered as consisting of canonical subensembles
with different particle numbers N . The grand canonical state energies and partition function
contain an additional chemical potential term. With this additional term the results obtained
for the canonical ensemble apply to the grand canonical ensemble, too.

The partition function for the grand canonical ensemble is given by


X P
Zgc (, V, T ) = e( k Ni,k k i )/kB T , (4.38)
i

whereas the probability distribution over the levels and particle numbers is
P
e( k Ni,k k i )/kB T
Pi = . (4.39)
Zgc

Note that the index range i is much larger than for a canonical ensemble, because each microstate
is now characterized by a set of particle numbers Ni,k , where k runs over the components.

R At this point we are in conflict with the notation that we used in the lecture course
on phenomenological thermodynamics (PC I http://www.epr.ethz.ch/education/
Thermodynamics_PCI/). In that course we defined the chemical potential as a molar
quantity, here it is a molecular quantity. The relation is PCI = NAv PCVI . Using the PC
I notation in the current lecture notes would be confusing in other ways, as is generally
used in statistical thermodynamics for the molecular chemical potential. A similar remark
applies to capital letters for state functions. Capital letters denote either a molecular
quantity or a molar quantity. The difference will be clear from the context. We note that in
general small letters for state functions (except for pressure p) denote extensive quantities
and capital letters (except for volume V ) denote intensive quantities.
Swendsens Postulates of Thermodynamics
Cautionary Remarks on Entropy
Swendsens Postulates
Entropy in Phenomenological Thermodynamics
Boltzmanns Entropy Definition
The Relation of State Functions to the Partition
Function
Entropy and the Partition Function
Helmholtz Free Energy
Gibbs Free Energy, Enthalpy, and Pressure
Irreversibility
Historical Discussion
Irreversibility as an Approximation
Entropy and Information
Gibbs Entropy
Von Neumann Entropy
Shannon Entropy

5 Entropy

5.1 Swendsens Postulates of Thermodynamics


5.1.1 Cautionary Remarks on Entropy
You will search in vain for a mathematical derivation or clear condensed explanation of entropy in
textbooks and textbook chapters on statistical thermodynamics (please tell me if you have found
one). There is a simple reason for it: no such derivation or explanation exists. With entropy
being a central concept, probably the central concept of the theory, this may appear very strange.
However, the situation is not as bad as it may appear. The theory and the expressions that can be
derived work quite well and have predictive power. There are definitions of entropy in statistical
thermodynamics (unfortunately, more than one) and they make some sense. Hence, while it may
be unnerving that we cannot derive the central state function from scratch, we can still do many
useful things and gain some understanding.
Textbooks tend to sweep the problem under the rug. We wont do that here. We try to make
an honest attempt to clarify what we do know and what we dont know about entropy before
accepting one working definition and base the rest of theory on this definition. It is probably best
to start with a set of postulates that explains what we expect from the quantity that we want to
define.

5.1.2 Swendsens Postulates


The following postulates are introduced and shortly discussed in Section 9.6 of Swendsens book
[Swe12]. We copy the long form of these postulates verbatim with very small alterations that
improve consistency or simplify the expression.
1. There exist equilibrium states of a macroscopic system that are characterized uniquely by a
small number of extensive variables.
2. The values assumed at equilibrium by the extensive variables of an isolated system in
the absence of internal constraints are those that maximize the entropy over the set of all
constrained macroscopic states.
3. The entropy of a composite system is additive over the constituent subsystems.
4. For equilibrium states the entropy is a monotonically increasing function of the energy.
5. The entropy is a continuous and differentiable function of the extensive variables.
We have omitted Swendsens last postulate (The entropy is an extensive function of the
extensive variables), because, strictly speaking, it is superfluous. If the more general third
42 Entropy

postulate of additivity is fulfilled, entropy is necessarily an extensive property.


Swendsens first postulate (Equilibrium States) establishes the formalism of thermodynamics,
while all the remaining postulates constitute a wish list for the quantity entropy that we need
to predict the equilibrium states. They are a wish list in the sense that we cannot prove that a
quantity with all these properties must exist. We can, however, test any proposed definition of
entropy against these postulates.
Some points need explanation. First, the set of postulates defines entropy as a state function,
although this may be hidden. The first postulate implies that in equilibrium thermodynamics
some extensive variables are state functions and that a small set of such state functions completely
specifies all the knowledge that we can have about a macroscopic system. Because entropy in turn
specifies the other state functions for an isolated system at equilibrium, according to the second
postulate (Entropy Maximization), it must be a state function itself. It must be an extensive
state function because of the third postulate (Additivity), but the third postulate requires more,
namely that entropies can be added not only for subsystems of the same type in the same state,
but also for entirely different systems. This is required if we want to compute a new equilibrium
state (or entropy change) after unifying different systems. Otherwise, the simple calorimetry
experiment of equilibrating a hot piece of copper with a colder water bath would already be
outside our theory. The fourth postulate (Monotonicity) is new compared to what we discussed in
phenomenological thermodynamics. For a classical ideal gas this postulate can be shown to hold.
This postulate is needed because it ensures that temperature is positive. The fifth postulate is a
matter of mathematical convenience, although it may come as a surprise in a theory based on
integer numbers of particles. We assume, as at many other points, that the system is sufficiently
large for neglecting any errors that arise from treating particle number as a real rather than an
integer number. In other words, these errors must be smaller than the best precision that we
can achieve in experiments. As we already know from phenomenological thermodynamics, the
fifth postulate does not apply to first-order phase transitions, where entropy has a discontinuity.
We further note that the second postulate is an alternative way of writing the Second Law of
Thermodynamics. The term in the absence of internal constraints in the second postulate
ensures that the whole state space (or, for systems fully described by Hamiltonian equations of
motion, the whole phase space) is accessible.

5.1.3 Entropy in Phenomenological Thermodynamics


Textbook authors are generally much more comfortable in discussing entropy as an abstract state
function in phenomenological thermodynamics than in discussing its statistical thermodynamics
aspects. We recall that the concept of entropy is not unproblematic in phenomenological
thermodynamics either. We had accepted the definition of Clausius entropy,
dqrev
ds = , (5.1)
T
where dqrev is the differentially exchanged heat for a reversible process that leads to the same
differential change in other state variables as an irreversible process under consideration and
T is the temperature. We could then show that entropy is a state function (Carnot process and
its generalization) and relate entropy via its total differential to other state functions. With
this definition we could further show that for closed systems, which can exchange heat, but
not volume work with their environment (dV = 0), minimization of Helmholtz free energy
f = u T s provides the equilibrium state and that for closed systems at constant pressure
(dp = 0), minimization of Gibbs free energy g = h T s provides the equilibrium state. Partial
molar Gibbs free energy is the chemical potential k,molar and via k,molecular = k,molar /NAv
it is related to terms in the partition function of the grand canonical ensemble, where we have
abbreviated k,molecular as k (Section 4.4).
5.1 Swendsens Postulates of Thermodynamics 43

We were unable in phenomenological thermodynamics to prove that the definition given in


Eq. (5.1) ensures fulfillment of the Second Law. We were able to give plausibility arguments
why such a quantity should increase in some spontaneous processes, but not more.

5.1.4 Boltzmanns Entropy Definition


Boltzmann provided the first statistical definition of entropy, by noting that it is the logarithm of
probability, up to a multiplicative and an additive constant. The formula s = k ln W by Planck,
which expresses Boltzmanns definition, omits the additive constant. We shall soon see why.
We now go on to test Boltzmanns definition against Swendsens postulates. From probability
theory and considerations on ensembles we know that for a macroscopic system, probability
density distributions for an equilibrium state are sharply peaked at their maximum. In other
words, the macrostate with largest probability is such a good representative for the equilibrium
state that it serves to predict state variables with better accuracy that the precision of experimental
measurements. It follows strictly that any definition of entropy that fulfills Swendsens postulates
must make s a monotonously increasing function of probability density1 for an isolated system.
Why the logarithm? Let us express probability (for the moment discrete again) by the
measure of the statistical weights of macrostates. We consider the isothermal combination of two
independent systems A with entropies sA and sB to a total system with entropy s = sA + sB
[WF12]. The equation for total entropy is a direct consequence of Swendsens third postulate. On
combination, the statistical weights A and B multiply, since the subsystems are independent.
Hence, with the monotonously increasing function f () we must have
s = f () = f (A B ) = f (A ) + f (B ) . (5.2)
The only solutions of this functional equation are logarithm functions. What logarithm we choose
will only influence the multiplicative constant. Hence, we can write
s = k ln , (5.3)
where, for the moment, constant k is unknown. Boltzmanns possible additive constant must
vanish at this point, because with such a constant, the functional equation (5.2), which specifies
additivity of entropy, would not have a solution.
It is tempting to equate in Eq. (5.3) in the context of phase space problems with the volume
of phase space occupied by the system. Indeed, this concept is known as Gibbs entropy (see
Section 5.4.1). It is plausible, since the phase space volume specifies a statistical weight for a
continuous problem. No problem arises if Gibbs entropy is used for equilibrium states as it then
coincides with Boltzmann entropy. There exists a conceptual problem, however, if we consider
approach to equilibrium. The Liouville theorem (see Section 2.2.2) states that the volume in
phase space taken up by a system is a constant of motion.2. Hence, Gibbs entropy is a constant
of motion for an isolated system and the equilibrium state would be impossible to reach from any
non-equilibrium state, which would necessarily occupy a smaller phase space volume. This leads
to the following cautionary remark:

R Statistical thermodynamics, as we introduce it in this lecture course, does not describe


dynamics that leads from non-equilibrium to equilibrium states. Different equilibrium
states can be compared and the equilibrium state can be determined, but we have made a
number of assumptions that do not allow us to apply our expressions and concepts to non-
equilibrium states without further thought. Non-equilibrium statistical thermodynamics is
explicitly outside the scope of the theory that we present here.
1Boltzmann was thinking in terms of discrete probability theory. As we want to use continuous probability theory
here, we have made the transition from probability to probability density.
2The theorem relies on uniform distribution in this volume at some point in time, but it applies here, as we have
seen before that such uniform distribution in an energy shell is a feature of the equilibrium state of an isolated system.
44 Entropy

A conceptual complication with Boltzmanns definition is that one might expect s to be


maximal at equilibrium for a closed system, too, not only for an isolated system. In classical
thermodynamics we have seen, however, that the equilibrium condition for a closed system
is related to free energy. Broadly, we could say that for a closed system probability must be
maximized for the system and its environment together. Unfortunately, this cannot be done
mathematically as the environment is very large (in fact, for mathematical purposes infinite).
The solution to this problem lies in the treatment of the canonical ensemble (Section 4.3). In
that treatment we have seen that energy enters into the maximization problem via the boundary
condition of constant total energy of the system that specifies what exactly is meant by thermal
contact between the system and its environment. We can, therefore, conclude that Boltzmanns
entropy definition, as further specified in Eq. (5.3), fulfills those of Swendsens postulates that
we have already tested and that the core idea behind it, maximization of probability (density) at
equilibrium is consistent with our derivation of the partition function for a canonical ensemble at
thermal equilibrium. We can thus fix k in Eq. (5.3) by deriving s from the partition function.

5.2 The Relation of State Functions to the Partition Function


5.2.1 Entropy and the Partition Function
We recall that we already computed internal energy u and heat capacity cV at constant volume
from the system partition function z (Section 4.3.3). For a canonical system (V = const.), which
is by definition at thermal equilibrium (reversible), we can identify qrev in Eq. (5.1) with3

dqrev = cV dT . (5.4)

Definite integration with substitution of cV by Eq. (4.35) gives [WF12],


Z T Z T  
cV 1 2 ln z
s s0 = dT = kB T dT (5.5)
0 T 0 T T T V
Z T   2    
1 2 ln z ln z
= kB T + 2kB T dT (5.6)
0 T T 2 V T V
Z T  2  Z T 
ln z ln z
= kB T dT + 2kB dT . (5.7)
0 T 2 V 0 T V

Partial integration provides


  Z T  Z T 
ln z ln z ln z
s s0 = kB T kB dT + 2kB dT (5.8)
T V 0 T V 0 T V
 
ln z T
= kB T + kB ln z

(5.9)
T V 0
u
= + kB ln z kB (ln z)T =0 , (5.10)
T
where we have used Eq. (4.34) to substitute the first term on the right hand side of Eq. (5.9). If
we assume that lim u/T = 0, the entropy at an absolute temperature of zero can be identified as
T 0
s0 = kB (ln z)T =0 . If there are no degenerate ground states, s0 = 0 in agreement with Nernsts
3Purists of statistical thermodynamics will shudder, as we now rely on the entropy definition of phenomenological
thermodynamics. We hide the fact that we are incapable of a strict general derivation and just relate the new concepts
of statistical thermodynamics to the concepts of phenomenological thermodynamics. In effect, we show how state
functions of phenomenological thermodynamics must be computed if both Boltzmanns and Clausius entropy
definitions apply.
5.2 The Relation of State Functions to the Partition Function 45

theorem (Third Law of Thermodynamics), as will be discussed in Section 7.2.2. Thus, by


associating u = 0 with T = 0 we obtain
  
u ln z
s = + kB ln z = kB + ln z . (5.11)
T ln T V

We see that under the assumptions that we have made the entropy can be computed from the
partition function. In fact, there should be a unique mapping between the two quantities, as both
the partition function and the entropy are state functions and thus must be uniquely defined by
the state of the system.
We now proceed with computing constant k in the mathematical definition of Boltzmann
entropy, Eq. (5.3). By inserting Eq. (4.9) into Eq. (5.3) we have
r1
!
X
s = k N ln N Ni ln Ni . (5.12)
i=0

We have neglected the term r on the right-hand side of Eq. (4.9), as is permissible if the number
N of particles is much larger than the number r of energy levels. Furthermore, according to Eq.
(4.25) and the definition of the partition function, we have Ni = N ei /kB T /Z. Hence,
r1 i /kB T
" !#
X e ei /kB T
s = k N ln N N ln N (5.13)
Z Z
i=0
r1 i /kB T r1 i /kB T r1 i /kB T
" #
X e X e X e i
= k N ln N N ln N + N ln Z + N
Z Z Z kB T
i=0 i=0 i=0
(5.14)
" Pr1 #
i /kB T
N i=0 i e
= k N ln N N ln N + N ln Z + , (5.15)
kB T Z

where we have used the definition of the partition function of going from Eq. (5.14) to (5.15).
Using Eq. (4.32) for substitution in the last term on the right-hand side of Eq. (5.15), we find
 
u
s = k N ln Z + . (5.16)
kB T
Comparison of Eq. (5.16) with Eq. (5.11) gives two remarkable results. First, the multiplicative
constant k in Boltzmanns entropy definition can be identified as k = kB = R/NAv . Second, for
the system of N identical, distinguishable classical particles, we must have

zdist = Z N . (5.17)

In other words, the partition function of a system of N identical, distinguishable, non-interacting


particles is the N th power of the molecular partition function.
It turns out that Eq. (5.17) leads to a contradiction if we apply it to an ideal gas. Assume
that we partition the system into two subsystems with particle numbers Nsub = N/2. The
internal-energy dependent term in Eq. (5.16) obviously will not change during this partitioning.
For the partition-function dependent term we have N ln Z for the total system and 2(N/2) ln Z 0
for the sum of the two subsystems. The molecular partition function in the subsystems differs,
because volume available to an individual particle is only half as large as in the total system.
For the inverse process of unifying the two subsystems we would thus obtain a mixing entropy,
although the gases in the subsystems are the same. This appearance of a mixing entropy for two
46 Entropy

identical ideal gases is called the Gibbs paradox. The Gibbs paradox can be healed by treating
the particles as indistinguishable. This reduces the statistical weight by N ! for the total system
and by (N/2)! for each subsystem, which just offsets the volume effect. Hence, for an ideal gas
we have
1 N
zindist = Z . (5.18)
N!
It may appear artificial to treat classical particles as indistinguishable, because the trajectory of
each particle could, in principle, be followed if they adhere to classical mechanics equations of
motion, which we had assumed. Note, however, that we discuss a macrostate and that we have
explicitly assumed that we cannot have information on the microstates, i.e., on the trajectories.
In the macrostate picture, particles in an ideal gas are, indeed, indistinguishable. For an ideal
crystal, on the other hand, each particle could be individually addressed, for instance, by high
resolution microscopy. In this case, we need to use Eq. (5.17).

5.2.2 Helmholtz Free Energy


Helmholtz free energy (German: Freie Energie) f is defined as

f = u Ts . (5.19)

This equation has a simple interpretation. From phenomenological thermodynamics we know


that the equilibrium state of a closed systems corresponds to a minimum in free energy. Among
all macrostates with the same energy u at a given temperature T , the equilibrium state is the one
with maximum entropy s. Furthermore, using Eq. (5.11) we have

f = u T (u/T + kB ln z) (5.20)
= kB T ln z . (5.21)

We note that this value of f , which can be computed from only the canonical partition function
and temperature, corresponds to the global minimum over all macrostates. This is not surprising.
After all, the partition function was found in a maximization of the probability of the macrostate.

5.2.3 Gibbs Free Energy, Enthalpy, and Pressure


All ensembles that we have defined correspond to equilibrium states at constant volume. To make
predictions for processes at constant pressure or to compute enthalpies h = u + pV and Gibbs
free energies g = f + pV we need to compute pressure from the partition function. The simplest
way is to note that p = (f /V )T,n . With Eq. (5.21) it then follows that
 
ln z
p = kB T , (5.22)
V T

where we have skipped the lower index n indicating constant molar amount. This is permissible
for the canonical ensemble, where the number of particles is constant by definition. From Eq.
(5.22) it follows that
 
ln z
pV = kB T (5.23)
ln V T
and
    
ln z ln z
h = u + pV = kB T + . (5.24)
ln T V ln V T
5.3 Irreversibility 47

Connoisseurs will notice the beautiful symmetry of this equation.


With Eq. (5.22) we can also compute Gibbs free energy (German: freie Enthalpie),
   
ln z
g = f + pV = kB T ln z . (5.25)
ln V T

5.3 Irreversibility
5.3.1 Historical Discussion
Daily experience tells us that some processes are irreversible. Phenomenological thermodynamics
had provided recipes for recognizing such processes by an increase in entropy for an isolated
system or decrease of free energy for a closed system. When Boltzmann suggested a link between
classical mechanics of molecules on a microscopic level and irreversibility of processes on the
macroscopic level, many physicists were irritated nevertheless. In retrospect it is probably fair to
say that a controversial discussion of Boltzmanns result could only ensue because atomistic or
molecular theory of matter was not yet universally accepted at the time. It is harder to understand
why this discussion is still going on in textbooks. Probably this is related to the fact that physicists
in the second half of the 19th and first half of the 20th believed that pure physics has implications
in philosophy, beyond the obvious ones in epistemology applied to experiments in the sciences.
If statistical mechanics is used to predict the future of the universe into infinite times, problems
ensue. If statistical mechanics is properly applied to well-defined experiments there are no such
problems.
Classical mechanics of particles does not involve irreversibility. The equations of motion
have time reversal symmetry and the same applies to quantum-mechanical equations of motion.
If the sign of the Hamiltonian can be inverted, the system will evolve backwards along the
same trajectory in phase space (or state space) that it followed to the point of inversion. This
argument is called Umkehreinwand or Loschmidt paradox and was brought up (in its classical
form) by Loschmidt. The argument can be refined and is then known as the central paradox:
Each microstate can be assigned a time-reversed state that evolves, under the same Hamiltonian,
backwards along the same trajectory. The two states should have the same probability. The
central paradox confuses equilibrium and non-equilibrium dynamics. At equilibrium a state and
the corresponding time-reversed state indeed have the same probability, which explains that the
macrostate of the system does not change and why processes that can be approximated by a series
of equilibrium states are reversible. If, on the other hand, we are not at equilibrium, there is no
reason for assuming that the probabilities of any two microstates are related. The system is at
some initial condition with a given set of probabilities and we are not allowed to pose symmetry
requirements to this initial condition.
The original Umkehreinwand, which is based on sign inversion of the Hamiltonian rather then
the momenta of microstates, is more serious than the central paradox. Time-reversal experiments
of this type can be performed, for instance, echo experiments in magnetic resonance spectroscopy
and optical spectroscopy. In some of these echo experiments, indeed the Hamiltonian is sign-
inverted, in most of these experiments application of a perturbation Hamiltonian for a short
time (pulse experiment) causes sign inversion of the density matrix. Indeed, the first paper on
observation of such a spin echo by Erwin Hahn was initially rejected with the argument that
he could not have observed what he claimed, as this would have violated the Second Law of
Thermodynamics. A macroscopic time-reversal experiment that creates a colorant echo in
corn syrup can be based on laminar flow [UNM12]. We note here that all these time-reversal
experiments are based on preparing a system in a non-equilibrium state. To analyze them, changes
in entropy or Helmholtz free energy must be considered during the evolution that can be reversed.
These experiments do not touch the question whether or not the same system will irreversibly
48 Entropy

approach an equilibrium state if left to itself for a sufficiently long time. We can see this easily for
the experiment with colorants and corn syrup. If, after setup of the initial state and evolution to
the point of time reversal, a long time would pass, the colorant echo would no longer be observed,
because diffusion of the colorants in corn syrup would destroy spatial correlation. The echo
relies on the fact that diffusion of the colorants in corn syrup can be neglected on the time scale
of the experiment, i.e., that equilibrium cannot be reached. The same is true for the spin echo
experiment, which fails if the evolution time is much longer than the transverse relaxation time
of the spins.
Another argument against irreversibility was raised by Zermelo, based on a theorem by
Poincar. The theorem states that any isolated classical system will return repeatedly to a
point in phase space that is arbitrarily close to the starting point. This argument is known as
Wiederkehreinwand or Zermelo paradox. We note that such quasi-periodicity is compatible with
the probability density formalism of statistical mechanics. The probability density distribution is
very sharply peaked at the equilibrium state, but it is not zero at the starting point in phase space.
The system fluctuates around the equilibrium state and, because the distribution is sharply peaked,
these fluctuations are very small most of the time. Once in a while the fluctuation is sufficiently
large to revisit even a very improbable starting point in phase space, but for a macroscopic system
this while is much longer than the lifetime of our galaxy. For practical purposes such large
fluctuations can be safely neglected, because they occur so rarely. That a system will never
evolve far from the equilibrium state once it had attained equilibrium is an approximation, but the
approximation is better than many other approximations that we use in physics. The statistical
error that we make is certainly much smaller than our measurement errors.

5.3.2 Irreversibility as an Approximation


If the whole of phase space is accessible the system will always tend to evolve from a less probable
macrostate to a more probable macrostate, until it has reached the most probable macrostate,
which is the equilibrium state. Equilibrium is dynamic. The microstate of each individual system
evolves in time. However, for most microstates the values of all state variables are the same as for
equilibrium within experimental uncertainty. In fact, the fraction of such microstates does not
significantly differ from unity. Hence, a system that has attained equilibrium once will be found
at equilibrium henceforth, as long as none of the external parameters is changed on which the
probability density distribution in phase space depends. In that sense, processes that run from a
non-equilibrium state to an equilibrium state are irreversible.
We should note at this point that all our considerations in this lecture course assume systems
under thermodynamic control. If microstate dynamics in phase space is slow compared to the
time scale of the experiment or simulation, the equilibrium state may not be reached. This may
also happen if dynamics is fast in the part of phase space where the initial state resides but
exchange dynamics is too slow between this part of phase space and the part of phase space
where maximum probability density is located.

5.4 Entropy and Information


5.4.1 Gibbs Entropy
For a system with a countable number of microstates an ensemble entropy can be defined by a
weighted sum over entropies of all microstates that are in turn expressed as kB ln Pi , which is
analogous to Boltzmanns entropy definition for a macrostate.
X
S = kB Pi ln Pi . (5.26)
i
5.4 Entropy and Information 49

This is the definition of Gibbs entropy, while Boltzmann entropy is assigned to an individual
microstate. Note that we have used a capital S because Gibbs entropy is a molecular entropy.
Using Eq. (4.25), we obtain for the system entropy s = N S,
X  i 
s = kB N Pi ln Z (5.27)
kB T
i
u
= + kB ln z , (5.28)
T
where we have assumed distinguishable particles, so that ln z = N ln Z. We have recovered Eq.
(5.11) that we had derived for the system entropy starting from Boltzmann entropy and assuming
a canonical ensemble. For a canonical ensemble of distinguishable particles, either concept can
be used. As noted above, Gibbs entropy leads to the paradox of a positive mixing entropy for
combination of two subsystems made up by the same ideal gas. More generally, Gibbs entropy is
not extensive if the particles are indistinguishable. The problem can be solved by redefining the
system partition function as in Eq. (5.18).
This problem suggests that entropy is related to the information we have on the system.
Consider mixing of 13 CO2 with 12 CO2 .4 At a time when nuclear isotopes were unknown, the
two gases could not be distinguished and mixing entropy was zero. With a sufficiently sensitive
spectrometer we could nowadays observe the mixing process by 13 C NMR. We will observe
spontaneous mixing. Quite obviously, the mixing entropy is not zero anymore.
This paradox cautions against philosophical interpretation of entropy. Entropy is a quantity
that can be used for predicting the outcome of physical experiments. It presumes an observer and
depends on the information that the observer has or can obtain.5 Statistical mechanics provides
general recipes for defining entropy, but the details of a proper definition depend on experimental
context.
Unlike the system entropy derived from Boltzmann entropy via the canonical ensemble,
Gibbs entropy is, in principle, defined for non-equilibrium states. Because it is based on the same
probability concept, Gibbs entropy in an isolated system is smaller for non-equilibrium states
than for equilibrium states.

5.4.2 Von Neumann Entropy


The concept of Gibbs entropy for a countable set of discrete states and their probabilities is easily
extended to continuous phase space and probability densities. This leads to the von Neumann
entropy,

S = kB Trace { ln } , (5.29)

where is the density matrix. Some physics textbooks dont distinguish von Neumann entropy
from Gibbs entropy [Sch06]. Von Neumann entropy is a constant of motion if an ensemble of
classical systems evolves according to the Liouville equation or a quantum mechanical system
evolves according to the Liouville-von-Neumann equation. It cannot describe the approach of an
isolated system to equilibrium. Coupling of the quantum mechanical system to an environment
can be described by the stochastic Liouville equation[Kub63; Tan06; VF74]
b i hb i b
= H, b + beq ) ,
b (b (5.30)
t ~
4This thought experiment was suggested to me by Roland Riek.
5One can speculate on philosophical interpretations. Irreversibility could be a consequence of partitioning the
universe into an observer and all the rest, a notion that resonates with intuitions of some mystical thinkers across
different religious traditions. Although the idea is appealing, it cannot be rationally proved. Rational thought already
implies that an observer exists.
50 Entropy

where b is a Markovian operator and eq the density matrix at equilibrium. This equation of
b
motion can describe quantum dissipative systems, i.e., the approach to equilibrium, without
relying explicitly on the concept of entropy, except for the computation of eq , which relies
on generalization of the Boltzmann distribution (see Section 6.1.2). However, to derive the
Markovian operator ,
b
b explicit assumptions on the coupling between the quantum mechanical
system and its environment must be made, which is beyond the scope of this lecture course.

5.4.3 Shannon Entropy


The concept of entropy has also been introduced into information theory. For any discrete random
number that can take values aj with probabilities P (aj ), the Shannon entropy is defined as
X
HShannon (a) = P (aj ) log2 P (aj ) . (5.31)
j

A logarithm to the basis of 2 is used here as the information is assumed to be coded by binary
numbers. Unlike for discrete states in statistical mechanics, an event may be in the set but still
have a probability P (aj ) = 0. In such cases, P (aj ) log2 P (aj ) is set to zero. Shannon entropy
is the larger the more random the distribution is, or, more precisely, the closer the distribution
is to a uniform distribution. Information is considered as deviation from a random stream of
numbers or characters. The higher the information content is, the lower the entropy.
Shannon entropy can be related to reduced Gibbs entropy = S/kB . It is the amount of
Shannon information that is required to specify the microstate of the system if the macrostate
is known. When expressed with the binary logarithm, this amount of Shannon information
specifies the number of yes/no questions that would have to be answered to specify the microstate.
We note that this is exactly the type of experiment presumed in the second Penrose postulate
(Section 2.3.1). The more microstates are consistent with the observed macrostate, the larger is
this number of questions and the larger are Shannon and Gibbs entropy. The concept applies
to non-equilibrium states as well as to equilibrium states. It follows, what was stated before
Shannon by G. N. Lewis: "Gain in entropy always means loss of information, and nothing more".
The equilibrium state is the macrostate that lacks most information on the underlying microstate.
We can further associate order with information, as any ordered arrangement of objects
contains information on how they are ordered. In that sense, loss of order is loss of information
and increase of disorder is an increase in entropy. The link arises via probability, as the total
number of arrangements is much larger than the number of arrangements that conform to a
certain order principle. Nevertheless, the association of entropy with disorder is only colloquial,
because in most cases we do not have quantitative descriptions of order.
Quantum Canonical Ensemble
Density Matrix
Quantum Partition Function
Quantum and Classical Statistics
-H/kBT
e
Types of Permutation Symmetry
Bose-Einstein Statistics
Fermi-Dirac Statistics
Maxwell-Boltzmann Statistics req =
Simple Quantum Systems
Spin S = 1/2
Harmonic Oscillator
Einstein and Debye Models of a Crystal
Tr{e -H/k T
} B

6 Quantum Ensembles

6.1 Quantum Canonical Ensemble


6.1.1 Density Matrix
We have occasionally referred to the quantum-mechanical density matrix in previous sections.
Before we discuss quantum ensembles, we need to fully specify this concept.

Concept 6.1.1 Density Matrix. The microstates that can be assumed by a system in a quantum
ensemble are specified by a possible set of wavefunctions i (i = 1 . . . r). The probability or
population of the ith microstate is denoted as Pi , and for the continuous case the probability
density for a given wavefunction is denoted as p(). The density operator is then given by
r1
X
b = P (i) |i i hi | (discrete) (6.1)
i=0
Z
b = p() |i i hi | (continuous) . (6.2)

The density operator can be expressed as a density matrix with respect to a set of basis
functions |ki. For exact computations the basis functions must form a countable complete
set that allows for expressing the system wavefunctions i as linear combinations of basis
functions. For approximative computations, it suffices that this linear combination is a good
approximation. The matrix elements of the density matrix are then given by
r1
X
nm = P (i) hm |i i hi | ni (discrete) (6.3)
Zi=0
nm = p() hm |i i hi | ni (continuous) . (6.4)

P
With the complex coefficients ck in the linear combination representation |i = k ck |ki, the
matrix elements are

nm = cm cn , (6.5)
52 Quantum Ensembles

where the asterisk denotes the complex conjugate and the bar for once denotes the ensemble
average. It follows that diagonal elements (m = n) are necessarily real, nn = |cn |2 and that
mn is the complex conjugate of nm . Therefore, the density matrix is Hermitian and the
density operator is self-adjoint. The matrix dimension is the number of basis functions. It is
often convenient to use the eigenfunctions of the system Hamiltonian H
b as the basis functions,
but the concept of the density matrix is not limited to this choice.

R That the density matrix can be expressed in the basis of eigenstates does not imply that
the ensemble can be represented as consisting of only eigenstates, as erroneously stated
by Swendsen [Swe12]. Off-diagonal elements of the density matrix denote coherent
superpositions of eigenstates, or short coherences. This is not apparent in Swendsens
simple example where coherence is averaged to zero by construction. The ensemble can
be represented as consisting of only eigenstates if coherence is absent. In that case the
density matrix is diagonal in the eigenbasis. Diagonal elements of the density matrix
denote populations of basis states.

In quantum mechanics, it is well defined what information we can have about the macrostate
of a system, because quantum measurements are probabilistic even for a microstate. We can
observe only quantities that are quantum-mechanical observables andDthese E observables are
represented by operators A. It can be shown that the expectation value A of any observable
b b
can be computed from the density matrix by [SJ01; Swe12]
D E n o
Ab = Trace bA b , (6.6)

where we have used operator notation for b to point out that b and A b must be expressed in the
same basis.
Since the expectation values of all observables are the full information that we can have
on a quantum system, the density matrix specifies the full information that we can have on
the ensemble. However, the density matrix does not fully specify the ensemble itself, i.e., we
cannot infer the probabilities P (i) or probability density function () from the density matrix
(Swendsen gives a simple example [Swe12]). This is another example for the information loss on
microstates that comes about when we can only observe macrostates and that is conceptually
equivalent to entropy. The von-Neumann entropy can be computed from the density matrix by
Eq. (5.29).
We note that there is one important distinction between classical and quantum-mechanical
observations for an individual system. In the quantum case we can specify only an expectation
value, and the second and third Penrose postulates (Section 2.3.1) do not apply: neither can
we simultaneously measure all observables (they may be incompatible), nor is the outcome of
a later measurement independent of the current measurement. However, quantum uncertainty
is much smaller than measurement errors for the large ensembles that we treat by statistical
thermodynamics. Hence, the Penrose postulates apply to the quantum-mechanical ensembles
that represent macrostates, although they do not apply to the microstates.
If all systems in a quantum ensemble populate the same microstate, i.e., they correspond to
the same wavefunction, the ensemble is said to be in a pure state. A pure state corresponds to
minimum rather than maximum entropy. Otherwise the system is said to be in a mixed state.

6.1.2 Quantum Partition Function


Energy quantization leads to a difficulty in using the microcanonical ensemble. The difficulty
arises because the microcanonical ensemble requires constant energy, which restricts our abilities
to assign probabilities in a set of discrete energy levels. However, as we derived the Boltzmann
6.1 Quantum Canonical Ensemble 53

distribution, partition function, entropy and all other state functions for classical systems from
the canonical ensemble anyway, we can simply ignore this problem. The canonical ensemble is
considered to be at thermal equilibrium with a heat bath (environment) of infinite size. It does
not matter whether this heat bath is of classical or quantum mechanical nature. For an infinitely
sized quantum system, the energy spectrum is continuous, which allows us to exchange energy
between the bath and any constituent system of the quantum canonical ensemble at will.
We can derive Boltzmann distribution and partition function for the density matrix by analogy
to the classical case. For that we consider the density matrix in the eigenbasis. The energies of the
eigenstates are the eigenvalues i of the Hamiltonian H. All arguments and mathematical steps
from Section 4.3.1 still apply, with a single exception: Quantum mechanics allows for microstates
that are coherent superpositions of eigenstates. The classical derivation carries over if and only
if we can be sure that the equilibrium density matrix can be expressed without contributions
from such microstates, which would lead to off-diagonal elements in the representation in the
eigenbasis of H. b This argument can indeed be made. Any superposition of two eigenstates
|ni and |mi with amplitudes |cn | and |cm | can be realized with arbitrary phase difference
between the two eigenfunctions. The microstates with the same |cn | and |cm | but different all
have the same energy. The entropy of a subensemble that populates these microstates is maximal
if the distribution of phase differences is uniform in the interval [0, 2). In that case cm cn
vanishes, i.e., such subensembles will not contribute off-diagonal elements to the equilibrium
density matrix.
We can now arrange the ei /kB T in matrix form,

= eH/kB T ,
b
(6.7)

with the matrix elements ii = ei /kB T and ij = 0 for i 6= j. The partition function is the sum
of all the diagonal elements of this matrix, i.e. the trace of . Hence,

eH/kB T
b
beq = n o, (6.8)
Trace eH/k
b BT

where we have used operator notation. This implies that Eq. (6.8) can be evaluated in any basis,
b In a different basis, ei /kB T needs to be computed as a matrix
not only the eigenbasis of H.
exponential and, in general, the density matrix eq will have non-zero off-diagonal elements in
such a different basis.
The quantum-mechanical partition function,
n o
Z = Trace eH/kB T ,
b
(6.9)

is independent of the choice of basis, as the trace of a matrix is invariant under unitary
transformations. Note that we have used a capital Z for a molecular partition function. This is
appropriate, as the trace of beq in Eq. (6.8) is unity. In the eigenbasis, the diagonal elements of
eq are the populations of the eigenstates at thermal equilibrium. There is no coherence for a
sufficiently large quantum ensemble at thermal equilibrium.
We note that the density matrix at thermal equilibrium can be derived in a more strict manner
by explicitly considering a system that includes both the canonical ensemble and the heat bath
and by either tracing out the degrees of freedom of the heat bath [Sch06] or relying on a series
expansion that reduces to only two terms in the limit of an infinite heat bath [Sch06; Swe12].
When approaching zero absolute temperature, the matrix element of in the eigenbasis that
corresponds to the lowest energy i becomes much larger than all the others. At T = 0, the
corresponding ground state is exclusively populated and the ensemble is in a pure state if there is
54 Quantum Ensembles

just one state with this energy. For T on the other hand, differences between the diagonal
matrix elements vanish and all states are equally populated. The ensemble is in a maximally
mixed state.

6.2 Quantum and Classical Statistics


6.2.1 Types of Permutation Symmetry
Classical particles are either distinguishable or non-distinguishable, a difference that influences
the relation between the system partition function and the molecular partition function (Section
5.2.1). Quantum particles are special. They are always indistinguishable, but there exist two
types that behave differently when two particles are permuted. For bosons, the wavefunction is
unchanged on such permutation, whereas for fermions the wavefunction changes sign. This sign
change does not make the particles distinguishable, as absolute phase of the wavefunction does
not correspond to an observable. However, it has important consequences for the population of
microstates. Two (or more) bosons can occupy the same energy level. In the limit T 0 they
will all occupy the ground state and form a Bose-Einstein condensate. Bosons are particles with
integer spin, with the composite boson 4 He (two protons, two neutrons, two electrons) probably
being the most famous example. In contrast, two fermions (particles with half-integer spin)
cannot occupy the same state, a fact that is known as Pauli exclusion principle. Protons, neutrons,
and electrons are fermions (spin 1/2), whereas photons are bosons (spin 1).
This difference in permutation symmetry influences the distribution of particles over energy
levels. The simplest example is the distribution of two particles to two energy levels l (for left)
and r (for right) [FR03]. For distinguishable classical particles four possible configurations
exist:
1. l is doubly occupied
2. l is occupied by particle A and r is occupied by particle B
3. l is occupied by particle B and r is occupied by particle A
4. r is doubly occupied.
For bosons and for indistinguishable classical particles as well, the second and third configuration
above cannot be distinguished. Only three configurations exist:
1. l is doubly occupied
2. l is occupied by one particle and r is occupied by one particle
3. r is doubly occupied.
For fermions, the first and third configuration of the boson case are excluded by the Pauli principle.
Only one configuration is left:
1. l is occupied by one particle and r is occupied by one particle.
Since the number of configurations enters into all probability considerations, we shall find
different probability distributions for systems composed of bosons, fermions, or distinguishable
classical particles. The situation is most transparent for an ideal gas, i.e. N non-interacting point
particles that have only translational degrees of freedom [WF12]. For such a system the spectrum
of energy levels is continuous.

6.2.2 Bose-Einstein Statistics


We want to derive the probability distribution for the occupation of energy levels by bosons. To
that end, we first pose the question how many configurations exist for distributing Ni particles
to Ai energy levels in the interval between i and i + d. Each level can be occupied by an
arbitrary number of particles. We picture the problem as a common set of particles and levels
that has Ni + Ai elements. Now we consider all permutations in this set and use the convention
that particles that stand left from a level are assigned to this level. For instance, the permutation
6.2 Quantum and Classical Statistics 55

{P1 , P2 , A1 , P3 , A2 , A3 } for three particles and three levels denotes a state where level A1 s
occupied by particles P1 and P2 , level A2 is occupied by particle P3 and level A3 is empty.
With this convention the last energy level is necessarily the last element of the set (any particle
standing right from it would not have an associated level), hence only (Ni + Ai 1)! such
permutations exist. Each permutation also encodes a sequence of particles, but the particles are
indistinguishable. Thus we have to divide by Ni ! in order to not double count configurations that
we cannot distinguish. We also have to divide by (Ai 1)!, since sets with a different sequence
of the levels but the same occupation numbers of each individual level are not distinguishable.
Alternatively, the division by (Ai 1)! can be understood by noting that the energy levels are
ordered by increasing energy and of all (Ai 1)! permutations of the lowest Ai 1 levels we
consider only the one that is properly ordered. Hence, the number of configurations and thus the
number of microstates in the interval between i and i + d is
(Ni + Ai 1)!
Ci = . (6.10)
Ni ! (Ai 1)!
The configurations in energy intervals with different indices i are independent of each other.
Hence, the statistical weight of a macrostate is
Y (Ni + Ai 1)!
= (6.11)
Ni ! (Ai 1)!
i
As the number of energy levels is, in practice, infinite, we can choose the Ai sufficiently large for
neglecting the 1 in Ai 1. In an exceedingly good approximation we can thus write
Y (Ni + Ai )!
= . (6.12)
Ni !Ai !
i
The next part of the derivation is the same as for the Boltzmann distribution in Section 4.3.1,
i.e., it relies on maximization of ln using
P the Stirling formula and considering the constraints
of conserved total particle number N = i Ni and conserved total energy of the system [WF12].
The initial result is of the form
Ni 1
= 
, (6.13)
Ai Be i 1
where B is related to the Lagrange multiplier by B = e and thus to the chemical potential
by B = e/(kB T ) . After a rather tedious derivation using the definitions of Boltzmann entropy
and (u/s)V = T we can identify with 1/kB T [WF12]. We refrain from reproducing this
derivation here, as the argument is circular: It uses the identification of k with kB in the definition
of Boltzmann entropy that we had made earlier on somewhat shaky grounds. We accept the
identification of || with 1/kB T as general for this type of derivations, so that we finally have
Ni 1
= . (6.14)
Ai Be B T 1
 i /k

Up to this point we have supposed nothing else than a continuous, or at least sufficiently
dense, energy spectrum and identical bosons. To identify B we must have information on this
energy spectrum and thus specify a concrete physical problem. When using the density of states
for an ideal gas consisting of quantum particles with mass m in a box with volume V (see Section
7.2 for derivation),
V
D() = 4 2 3 m3/2 1/2 , (6.15)
h
we find, for the special case Bei /kB T  1,
(2mkB T )3/2 V
B= . (6.16)
h3 N
56 Quantum Ensembles

6.2.3 Fermi-Dirac Statistics


The number Ni of fermions in an energy interval with Ai levels cannot exceed Ai . The number
of allowed configurations is now given by the number of possibilities to select Ni out of Ai levels
that are populated, whereas the remaining levels remain empty. As each level can exist in only
one of two conditions, populated or empty, this is a binomial distribution problem as we have
solved in Section 3.1.6. In Eq. (3.23) we need to substitute N by Ai and n by Ni . Hence, the
number of allowed configurations in the energy interval between i and i + i is given by

Ai !
Ci = (6.17)
Ni ! (Ai Ni )!

and, considering mutual independence of the configurations in the individual energy intervals,
the statistical weight of a macrostate for fermions is
Y Ai !
= . (6.18)
Ni ! (Ai Ni )!
i

Again, the next step of the derivation is analogous to derivation of the Boltzmann distribution
in Section 4.3.1 [WF12]. We find
Ni 1
= . (6.19)
Ai Be B T + 1
 i /k

For the special case Bei /kB T  1, B is again given by Eq. (6.16). Comparison of Eq. (6.19
) with Eq. (6.14) reveals as the only difference the sign of the additional number 1 in the
denominator on the right-hand side of the equations. In the regime Bei /kB T  1, for which we
have specified B, this difference is negligible.
It is therefore of interest when this regime applies. As i 0 in the ideal gas problem, we
have ei /kB T 1, so that B  1 is sufficient for the regime to apply. Wedler and Freund [WF12]
have computed values of B according to Eq. (6.16) for the lightest ideal gas, H2 , and have
found B  1 for p = 1 bar down to T = 20 K and at ambient temperature for pressures up to
p = 100 bar. For heavier molecules, B is larger under otherwise identical conditions. Whether a
gas atom or molecule is a composite boson or fermion thus does not matter, except at very low
temperatures and very high pressures. However, if conduction electrons in a metal, for instance
in sodium, are considered as a gas, their much lower mass and higher number density N/V leads
to B  1 at ambient temperature and even at temperatures as high as 1000 K. Therefore, a gas
model for conduction electrons (spin 1/2) must be set up with Fermi-Dirac statistics.

6.2.4 Maxwell-Boltzmann Statistics


In principle, atoms and molecules are quantum objects and not classical particles. This would
suggest that the kinetic theory of gases developed by Maxwell before the advent of quantum
mechanics is deficient. However, we have already seen that for particles as heavy as atoms and
molecules and number densities as low as in gases at atmospheric pressure or a bit higher, the
difference between Bose-Einstein and Fermi-Dirac statistics vanishes, unless temperature is very
low. This suggests that, perhaps, classical Maxwell-Boltzmann statistics is indeed adequate for
describing gases under common experimental conditions.
We assume distinguishable particles. Each of the Ni particles can be freely assigned to one
of the Ai energy levels. All these configurations can be distinguished from each other, as we can
picture each of the particles to have an individual tag. Therefore,

Ci = (Ai )Ni (6.20)


6.2 Quantum and Classical Statistics 57

configurations can be distinguished in the energy interval between i and i + i . Because the
particles are distinguishable (tagged), the configurations in the individual intervals are generally
not independent from each other, i.e. the total number of microstates does not factorize into
the individual numbers of microstates in the intervals. We obtain more configurations than that
because we have the additional choice of distributing the N tagged particles to r intervals. We
have already solved this problem in Section 4.3.1, the solution is Eq. (4.6). By considering the
additional number of choices, which enters multiplicatively, we find for the statistical weight of a
macrostate
N! Nr1
= AN0 N1
0 A1 . . . Ar1 (6.21)
N0 !N1 ! . . . Nr1 !
Y ANi
i
= N! . (6.22)
Ni !
i

It appears that we have assumed a countable number r of intervals, but as in the derivations
for the Bose-Einstein and Fermi-Dirac statistics, nothing prevents us from making the intervals
arbitrarily narrow and their number arbitrarily large.
Again, the next step in the derivation is analogous to derivation of the Boltzmann distribution
in Section 4.3.1 [WF12]. All the different statistics differ only in the expressions for , constrained
maximization of ln uses the same Lagrange ansatz. We end up with
Ni 1
= . (6.23)
Ai Be /kB T
 i

Comparison of Eq. (6.23) with Eqs. (6.14) and (6.19) reveals that, again, only the 1 in the
denominator on the right-hand side makes the difference, now it is missing. In the regime, where
Bose-Einstein and Fermi-Dirac statistics coincide to a good approximation, both of them also
coincide with Maxwell-Boltzmann statistics.
There exist two caveats. First, we already know that the assumption of distinguishable
particles leads to an artificial mixing entropy for two subsystems consisting of the same ideal gas
or, in other words, to entropy not being extensive. This problem does not, however, influence
the probability distribution, it only influences scaling of entropy with system size. We can solve
it by an ad hoc correction when computing the system partition function from the molecular
partition function. Second, to be consistent we should not use the previous expression for B,
because it was derived under explicit consideration of quantizationP of momentum.1 However, for
Maxwell-Boltzmann statistics B can be eliminated easily. With i Ni = N we have from Eq.
(6.23)
1 X
N= Ai ei /kB T , (6.24)
B
i

which gives
1 N
=P i /kB T
. (6.25)
B i Ai e

With this, we can express the distribution function as

Ni Ai ei /kB T
Pi = =P i /kB T
. (6.26)
N i Ai e

1This is more a matter of taste than of substance. As long as Bei /kN T  1, we can approximate any type
of quantum statistics by Maxwell-Boltzmann statistics before solving for B. We are thus permitted to freely mix
Maxwell-Boltzmann statistics with quantum-mechanical equations of motion.
58 Quantum Ensembles

Comparison of Eq. (6.26) with the Boltzmann distribution given by Eq. (4.25) reveals the
factors Ai as the only difference. Thus, the probability distribution for Maxwell-Boltzmann
statistics deviates from the most common form by the degree of degeneracy Ai of the individual
levels. This degeneracy entered the derivation because we assumed that within the intervals
between i and i + i several levels exist. If i is finite, we speak of near degeneracy. For
quantum systems, degeneracy of energy levels is a quite common phenomenon even in small
systems where the energy spectrum is discrete. In order to describe such systems, the influence
of degeneracy on the probability distribution must be taken into account.

Concept 6.2.1 Degeneracy. In quantum systems with discrete energy levels there may exist
gi quantum states with the same energy i that do not coincide in all their quantum numbers.
This phenomenon is called degeneracy and gi the degree of degeneracy. A set of gi degenerate
levels can be populated by up to gi fermions. In the regime, where Boltzmann statistics is
applicable to the quantum system, the probability distribution considering such degeneracy is
given by

Ni gi ei /kB T
Pi = =P i /kB T
(6.27)
N i gi e

and the molecular partition function by


X
Z= gi ei /kB T . (6.28)
i

The condition that degenerate levels do not coincide in all quantum numbers makes sure that the
Pauli exclusion principle does not prevent their simultaneous population with fermions.
At this point we can summarize the expected number of particles with chemical potential at
level i with energy i and arbitrary degeneracy gi for Bose-Einstein, Fermi-Dirac, and Boltzmann
statistics:
gi
Ni = ( )/(k T ) Bose Einstein statistics (6.29)
e i B 1
gi
Ni = ( )/(k T ) Fermi Dirac statistics (6.30)
e i B +1
gi
Ni = ( )/(k T ) Boltzmann statistics . (6.31)
e i B
P
Note that the chemical potential in these equations is determined by the condition N = i Ni .
The constant B in the derivations above is given by B = e/(kB T ) . If N is not constant, we
have = 0 and thus B = 1.

6.3 Simple Quantum Systems


6.3.1 Spin S = 1/2
The simplest quantum system is a two-level system and probably the best approximation to
isolated two-level systems is found in magnetic resonance spectroscopy of dilute S = 1/2 spin
systems. The Hamiltonian for an electron spin S = 1/2 in an external magnetic field along z is
given by
H
b = ~B0 Sbz , (6.32)
where = gB /~ is the gyromagnetic ratio and B0 is the magnetic field expressed in units of 1
Tesla. The two states are designated by magnetic spin quantum number mS = 1/2 and have
6.3 Simple Quantum Systems 59

energies  = ~B0 /2. The partition function is

Z = e~B0 /2kB T + e~B0 /2kB T , (6.33)

and the expectation value of Sbz , which is proportional to longitudinal magnetization, by


D E X
Sbz = mS P (mS ) (6.34)
(1/2) e~B0 /2kB T + (1/2)e~B0 /2kB T
= (6.35)
Z
1
= tanh (~B0 /2kB T ) . (6.36)
2
Usually one has ~B0  2kB T , which is called the high-temperature approximation. The series
expansion of the hyperbolic tangent,
1 2
tanh(x) = x x3 + x5 + . . . , (6.37)
3 15
can then be restricted to the leading term, which gives
D E ~B0
Sbz = . (6.38)
4kB T

6.3.2 Harmonic Oscillator


A diatomic molecule has one vibrational mode along the bond direction x. If we assign masses
mA and mB to the two atoms and a force constant f to the bond, we can write the Hamiltonian as

b = 1fx pb2
H b2 + , (6.39)
2 2
where the reduced mass is
mA mB
= (6.40)
mA + mB
and where the first term on the right-hand side of Eq. (6.39) corresponds to potential energy and
the second term to kinetic energy.
Eq. (6.39) can be cast in the form
2 2
b = 1 2 (R RE )2 ~ d ,
H (6.41)
2 2 dR2
where we have substituted x
b by the deviation of the atom-atom distance R from the bond length
RE and introduced the angular oscillation frequency of a classical oscillator with
s
f
= . (6.42)

Eq. (6.41) produces an infinite number of eigenstates with energies


 
1
v = ~ v + , (6.43)
2
where v = 0, 1, . . . , is the vibrational quantum number. All energies are positive, even the
one of the ground state with v = 0. This residual zero-point vibration can be considered as a
60 Quantum Ensembles

consequence of Heisenbergs uncertainty principle, since for a non-oscillating diatomic molecule


atom coordinates as well as momentum would be sharply defined, which would violate that
principle. In the context of statistical thermodynamics the unfortunate consequence is that for an
ensemble of N diatomic molecules for T 0 the vibrational contribution to the internal energy
u approaches u0 = N ~/2 and thus the term u/T in the entropy expression (5.11) approaches
infinity. We ignore this problem for the moment.
The partition function of the harmonic oscillator is an infinite series,

X
Z= e~(v+1/2)/kB T (6.44)
v=0

X
~/2kB T
=e e~v/kB T (6.45)
v=0
X  v
= e~/2kB T e~/kB T (6.46)
v=0
X
= e~/2kB T xn . (6.47)
n=0

where we have substituted x = e~/kB T and n =Pv to obtain the last line. Since for finite
temperatures 0 < e~/kB T < 1, the infinite series n
n=0 x converges to 1/(1 x). Hence,

e~/2kB T
Z= . (6.48)
1 e~/kB T
We can again discuss the behavior for T 0. In the denominator, the argument of the
exponential function approaches , so that the denominator approaches unity. In the numerator
the argument of the exponential function also approaches , so that the partition function
approaches zero and Helmholtz free energy f = kB T ln Z can only be computed as a limiting
value. The term kB ln Z in the entropy equation (5.11) approaches .
This problem can be healed by shifting the energy scale by E = ~/2. We then have2
1
Z= . (6.49)
1 e~/kB T
With this shift, the partition function and the population of the ground state v = 0 both approach
1 when the temperature approaches zero. For the term u/T in the entropy expression we still
need to consider a limiting value, but it can be shown that lim u/T = 0. Since kB ln Z = 0 for
T 0
Z = 1, entropy of an ensemble of harmonic oscillators vanishes at the zero point in agreement
with Nernsts theorem. Helmholtz free energy f = kB T ln Z approaches zero.

R For computing a Boltzmann distribution we can shift all energy levels by the same offset
E without influencing the P (i), as such a shift leads to a multiplication by the same
factor of the numerator and of all terms contributing to the partition function. Such a shift
can remove an infinity of the partition function.

This partition function can also be expressed with a characteristic vibrational temperature
~
vib = . (6.50)
kB
2The shift does not influence the denominator, as it merely removes the first factor on the right-hand side of Eq.
(6.45).
6.3 Simple Quantum Systems 61

This temperature is usually higher than room temperature. We have


1
Zvib = . (6.51)
1 evib /T
Thus, Z 1 at room temperature, which implies that only the vibrational ground state
is significantly populated. Vibration does not significantly contribute to entropy at room
temperature.

6.3.3 Einstein and Debye Models of a Crystal


The considerations on the harmonic oscillator can be extended to a simple model for vibrations in
a crystal. If we assume that all atoms except one are fixed at their average locations, the potential
at the unique atom is parabolic. This assumption made by Einstein may at first sight violate his
own dictum that "Everything should be made as simple as possible, but not simpler.". We shall
come back to this point below. For the moment we consider Einsteins approach as a very simple
mean field approach. Instead of the one-dimensional harmonic oscillator treated in Section 6.3.2,
we now have a three-dimensional harmonic oscillator. For sufficiently high point symmetry at the
unique atom, we can assume an isotropic force constant f . Each atom is then described by three
independent harmonic oscillators along three orthogonal directions. The harmonic oscillators of
different atoms are also independent by construction. Because we want to compute an absolute
internal energy we revert to the partition function of the harmonic oscillator without energy shift
given in Eq. (6.48). The partition function for a crystal with N atoms, considering that the atoms
in a crystal lattice are distinguishable and that thus Eq. (5.17) applies, is then given by
!3N
e~/2kB T
Z= . (6.52)
1 e~/kB T

Internal energy can be computed by Eq. (4.34) [WF12],


!3N
e ~/2k B T
uvib = kB T 2 ln (6.53)
T 1 e~/kB T
V
  
2 ~ 
~/kB T
= 3kB N T ln 1 e (6.54)
T 2kB T V
" #
~ 2
~/kB T e ~/k B T
= 3kB N T 2 2
+ ~/k T
(6.55)
2kB T 1e B

3 3N ~
= N ~ + ~/k T . (6.56)
2 e B 1
With the characteristic vibrational temperature vib introduced in Eq. (6.50) and by setting
N = NAv to obtain a molar quantity, we find
3 3Rvib
Uvib = Rvib + /T . (6.57)
2 e vib 1
The molar heat capacity of an Einstein solid is the derivative of Uvib with respect to T . We
note that we do not need to specify constant volume or constant pressure, since this simple model
depends on neither of these quantities. We find

(vib /T )2 evib /T
Cvib = 3R 2 . (6.58)
evib /T 1
62 Quantum Ensembles

According to the rule of Dulong and Petit we should obtain the value 3R for T . Since
the expression becomes indeterminate (0/0), we need to compute a limiting value, which is
possible with the approach of de lHospital where we separately differentiate the numerator and
denominator[WF12]. The derivation is lengthy but it indeed yields the limiting value 3R:

(vib /T )2 evib /T
lim Cvib = lim 3R 2 (6.59)
T T evib /T 1
2 (vib /T ) vib /T 2

= 3R lim   (6.60)
T 2 1 evib /T evib /T (vib /T 2 )
(vib /T )
= 3R lim  (6.61)
T 1 evib /T

vib /T 2

= 3R lim (6.62)
T evib /T (vib /T 2 )

= 3R . (6.63)

In Eq. (6.59) in the numerator and in going from Eq. (6.60) to (6.61) we have set evib /T to 1, as
we may for T . As the expression was still indeterminate, we have computed the derivatives
of numerator and denominator once again in going from Eq. (6.61) to (6.62) and finally we have
once more set evib /T to 1 in going from Eq. (6.62) to (6.63). We see that Einsteins very
simple model agrees with the rule of Dulong and Petit.

R The model of the Einstein solid differs from a model of NAv one-dimensional harmonic
oscillators according to Section 6.3.2 only by a power of 3 in the partition function, which,
after computing the logarithm, becomes a factor of 3 in the temperature-dependent term of
Uvib and thus in Cvib . Hence, in the high-temperature limit the vibrational contribution to
the molar heat capacity of a gas consisting of diatomic molecules is equal to R. It follows
that, in this limit, each molecule contributes an energy kB T to the internal energy, i.e.
each of the two degrees of freedom (potential and kinetic energy of the vibration) that are
quadratic in the coordinates contributes a term kB T /2. This agrees with the equipartition
theorem. Likewise, the Einstein solid agrees with this theorem.

From experiments it is known that molar heat capacity approaches zero when temperature
approaches zero. Again the limiting value can be computed by the approach of de l Hospital
[WF12], where this time we can neglect the 1 in evib /T 1, as evib /T tends to infinity for
T 0. In the last step we obtain
1
lim Cvib = 6R lim =0. (6.64)
T 0 T 0 evib /T
Thus, the Einstein solid also agrees with the limiting behavior of heat capacity at very low
temperatures.
Nevertheless the model is too simple, and Einstein was well aware of that. Vibrations of
the individual atoms are not independent, but rather collective. The lattice vibrations, called
phonons, have a spectrum whose computation is outside the scope of the Einstein model. A
model that can describe this spectrum has been developed by Debye based on the density of states
of frequencies . This density of states in turn has been derived by Rayleigh and Jeans based on
the idea that the phonons are a system of standing waves in the solid. It is given by [WF12]
4 2
D () = V . (6.65)
c3
Debye replaced c by a mean velocity of wave propagation in the solid, considered one longitudinal
and two transverse waves and only the 3N states with the lowest frequencies, as the solid has
6.3 Simple Quantum Systems 63

only 3N vibrational degrees of freedom. These considerations lead to a maximum phonon


frequency max and, after resubstitution of the mean velocity, to a frequency spectrum that is still
3 . Instead of the characteristic vibration temperature, it is
proportional to 2 and scales with max
now convenient to define the Debye temperature

hmax
D = . (6.66)
kB
In this model the molar heat capacity of the solid becomes
3 Z D /T
x4 e x

T
Cvib = 9R dx (6.67)
D 0 (ex 1)2

The integral can be evaluated numerically after series expansion and finally Debyes T 3 law,

T3
lim Cvib = 233.8R , (6.68)
T 0 3D

results. This law does not only correctly describe that the heat capacity vanishes at absolute zero,
it also correctly reproduces the scaling law, i.e., the T 3 dependence that is found experimentally.
The high-temperature limit can also be obtained by series expansion and is again Dulong-Petits
value of 3R.
The Debye model is still an approximation. Phonon spectra of crystalline solids are not
featureless. They are approximated, but not fully reproduced, by a 2 dependence. The deviations
from the Debye model depend on the specific crystal structure.
Separation of contributions
Collective Degrees of Freedom
Factorization of Energy Modes
Translational Partition Function
Density of States of an Ideal Gas
Partition Function and Accessible States
Nuclear Spin Partition Function
High-Temperature Limit
Symmetry Requirements
Rotational Partition Function
Rigid Rotor Assumption and Rotamers
Accessible States and Symmetry
Vibrational Partition Function
The Harmonic Oscillator Extended
Vibrational Contributions to U , CV , and S
Electronic Partition Function
Equilibrium Constant for Gas Reactions

7 Partition Functions of Gases

7.1 Separation of contributions


7.1.1 Collective Degrees of Freedom
In Section 6.3.3 we have seen that the treatment of condensed phases can be complicated by
collective motion of particles. Such effects are absent in an ideal gas that consists of point
particles, a model that is reasonable for noble gases far from condensation. For gases consisting
of molecules, it does not suffice to consider only translational motion as in Maxwells kinetic
theory of gases. We see this already when considering H2 gas, where each molecule can be
approximated by a harmonic oscillator (Section 6.3.2). Neglect of the vibrational degrees of
freedom will lead to wrong results for internal energy, heat capacity, the partition function, and
entropy, at least at high temperatures. In fact, an H2 molecule is not only an oscillator, it is also a
rotor. As a linear molecule it has two rotational degrees of freedom, which also contribute to
internal energy and to the partition function.
In principle, we could try to ignore all this and treat each atom as one particle. If the
Hamiltonian includes the potentials that characterize interaction between the particles, our
equations of motion would be correct. In practice, such a treatment is inconvenient and it is
better to group the spatial degrees of freedom according to the type of motion. The H2 molecule
has 3 translational degrees of freedom, 2 rotational degrees of freedom, and 1 vibrational degree
of freedom in the collective motion picture. The sum is 6, as expected for two atoms with each of
them having 3 translational degrees of freedom in an independent motion picture. In general, a
molecule with n atoms has fs = 3n spatial degrees of freedom, 3 of which are translational, 3
are rotational, except for linear molecules, which have only 2 rotation degrees of freedom, and
the rest are vibrational. We note that the number of degrees of freedom in phase space is f = 2fs
because each spatial degree of freedom is also assigned a momentum degree of freedom.
These considerations take care of particle motion. Further contributions to internal energy
and to the partition function can arise from spin. In both closed-shell and open-shell molecules,
nuclear spin can play a role. This is indeed the case for H2 , which can exist in ortho and para
states that differ in correlation of the nuclear spins of the two hydrogen atoms. For open-shell
molecules electron spin degrees of freedom must be considered. This is the case, for instance,
for O2 , which has a triplet ground state. In this case, rotational and spin degrees of freedom
correspond to similar energies and couple. Finally, at sufficiently high temperatures electronic
excitation becomes possible and then also makes a contribution to the partition function.
66 Partition Functions of Gases

In many cases, the individual contributions are separable, i.e. the modes corresponding to
different types of motions can be treated independently. Roughly speaking, this results from a
separation of energy ranges (frequency bands) of the modes and a corresponding separation of
time scales. Nuclear spin degrees of freedom have much lower energy than rotational degrees
of freedom which usually have much lower energy than vibrational degrees of freedom which
have much lower energies than electronic excitation. The independence of nuclear and electron
motion is basis of the Born-Oppenheimer approximation and the independence of rotational
and vibrational motion is invoked when treating a molecule as a rigid rotor. Separability of
energy modes leads to a sum rule for the energy contributions for a single closed-shell molecule
[Mac98],

j = j,trs + j,ns + j,rot + j,vib + j,el , (7.1)

where j,trs , j,ns , j,rot , j,vib , and j,el are the translational, nuclear spin, rotational, vibrational,
and electronic contributions, respectively. For a monoatomic molecule (atom) j,rot and j,vib
vanish. If both the number of neutrons and of protons in the nucleus is even, the nucleus has spin
I = 0. In that case the nuclear spin contribution vanishes for an atom, even in the presence of an
external magnetic field. If all nuclei have spin zero, the nuclear spin contribution also vanishes
for a diatomic or multi-atomic molecule.

7.1.2 Factorization of Energy Modes


If we assume the equipartition theorem to hold, or even more generally, the whole system to
attain thermal equilibrium, there must be some coupling between the different modes. If we say
that the energy modes are separable, we assume weak coupling, which means that for statistical
purposes we can assume the modes to be independent of each other. The consequence for the
computation of the partition function can be seen by considering a system of N particles with an
mode associated with quantum number k and an mode associated with quantum number r
[Mac98]. The total energy of a single molecule of this type is j = j,k + j,r The molecular
partition function is given by
XX
Z= e(k +r ) . (7.2)
k r

This sum can be rewritten as


XX
Z= ek er (7.3)
k r
X X
= ek er (7.4)
k r
X X
= ek Z = Z ek (7.5)
k k
= Z Z . (7.6)

We see that the total partition function is the product of the partition functions corresponding to
the individual modes. This consideration can be extended to multiple modes. With Eq. (7.1) it
follows that

Z = Ztrs Zns Zrot Zvib Zel . (7.7)

By considering Eq. (5.17) or Eq. (5.18) we see that we can also compute the partition function
for a given mode for all N particles before multiplying the modes. We have already seen that we
7.2 Translational Partition Function 67

N /N ! to heal the Gibbs paradox. What about the other, internal degrees of
must set ztrs = Ztrs
freedom? If two particles with different internal states are exchanged, they must be considered
distinguishable, exactly because their internal state tags them. Hence, for all the other modes
we have z = ZN . Thus,

1
z= (Ztrs Zns Zrot Zvib Zel )N (7.8)
N!
ZN
= trs Zns
N N
Zrot N
Zvib ZelN . (7.9)
N!
Accordingly, we can consider each of the partition functions in turn. We also note that separability
of the energies implies factorization of the molecular wavefunction,

= trs ns rot vib el (7.10)

7.2 Translational Partition Function


7.2.1 Density of States of an Ideal Gas
First, we derive the density of states that we had already used in computing the distribution
functions for quantum gases. We consider a quantum particle in a three-dimensional cubic box
with edge length a [WF12]. The energy is quantized with integer quantum numbers nx , ny , and
nz corresponding to the three pairwise orthogonal directions that span the cube,

h2 2 2 2

trs = n x + n y + n z (7.11)
8ma2
1
p2x + p2y + p2z .

= (7.12)
2m
It follows that momentum is also quantized with |pi | = (h/2a)ni (i = x, y, z). It is convenient
to consider momentum in a Cartesian frame where h/2a is the unit along the x, y, and z axes.
Each state characterized by a unique set of translational quantum numbers (nx , ny , nz ) owns a
small cube with volume h3 /8a3 in the octant with x 0, y 0, and z 0. Since momentum
can also be negative, we need to consider all eight octants, so that each state owns a cell in
momentum space with volume h3 /a3 . In order to go to phase space, we need to add the spatial
coordinates. The particle can move throughout the whole cube with volume a3 . Hence, each
state owns a phase space volume of h3 .
By rearranging Eq. (7.11) we can obtain an equation that must be fulfilled by the quantum
numbers,

n2x n2y n2z


a
2 + a
2 + a
2 = 1 (7.13)
h 8m h 8m h 8m

and by using Eq. (7.12) we can convert it to an equation that must be fulfilled by the components
of the momentum vector,

p2x p2y p2z


1
2 + 1
2 + 1
2 = 1 . (7.14)
2 8m 2 8m 2 8m

All states with quantum numbers that make the expression on the left-hand side of Eq. (7.13)
or Eq. (7.14) smaller than 1 correspond to energies that are smaller than . The momentum

associated with these states lies in the sphere defined by Eq. (7.14) with radius 12 8m and
68 Partition Functions of Gases

volume 6 (8m)3/2 . With cell size h3 /a3 in momentum space the number of cells with energies
smaller than  is

8 2 V
N () = (m)3/2 , (7.15)
3 h3

where we have substituted a3 by volume V of the box. The number of states in an energy interval
between  and  + d is the first derivative of N () with respect to  and is the sought density of
states,
V
D() = 4 2 3 m3/2 1/2 . (7.16)
h

7.2.2 Partition Function and Accessible States


This density of states is very high, so that we can replace the sum over the quantum numbers ni
in the partition function of the canonical ensemble by an integral [Mac98],
Z
2 2 /8ma2
Ztrs,i = eni h dni (i = x, y, z) (7.17)
o
r
2m a
= . (7.18)
h

The contributions along orthogonal spatial coordinates are also independent of each other and
factorize. Hence,
 3/2
2mkB T
Ztrs = Ztrs,x Ztrs,y Ztrs,z = V , (7.19)
h2

where we have again substituted a3 by V and, as by now usual, also by 1/kB T . The
corresponding molar partition function is
" 3/2 #NAv
1 2mkB T
ztrs = V . (7.20)
NAv ! h2

At this point it is useful to introduce another concept:

Concept 7.2.1 Number of accessible states. The molecular canonical partition function Z is
a measure for the number of states that are accessible to the molecule at a given temperature.

This can be easily seen when considering

Ni gi ei /kB T
Pi = = (7.21)
N Z

and i Pi = 1. If we consider a mole of 4 He (bosons) at 4.2 K, where it liquifies, we find


P
that Ztrs /NAv 7.5, which is not a large number [Mac98]. This indicates that we are close
to breakdown of the regime where Bose-Einstein statistics can be approximated by Boltzmann
statistics.

R For T 0 only the g0 lowest energy states are populated. In the absence of ground-state
degeneracy, g0 = 1, we find Z = 1 and with an energy scale where U (T = 0) = 0 we
have S(0) = 0 in agreement with Nernsts theorem.
7.3 Nuclear Spin Partition Function 69

An expression for the translational contribution to the entropy of an ideal gas can be derived
from Eq. (7.19), Eq. (5.18), and Eq. (5.11). We know that u = 3N kB T /2, so that we only need
to compute ln ztrs ,
1 N
ln ztrs = ln z (7.22)
N ! trs
= ln N ! + N ln ztrs (7.23)
= N ln N + N + N ln ztrs (7.24)
 ztrs 
= N 1 + ln , (7.25)
N
where we have used Stirlings formula to resolve the factorial. Thus we find
u
s = + kB ln z (7.26)
T
3  ztrs 
= N kB + kB N 1 + ln (7.27)
2   N
5 ztrs
= N kB + ln (7.28)
2 N
By using Eq. (7.19) we finally obtain the Sackur-Tetrode equation
( "  3 #)
5 2mkB T 2 V
s = N kB + ln . (7.29)
2 h2 N

To obtain the molar entropy Sm , N has to be replaced by NAv . Volume can be substituted by
pressure and temperature, by noting that the molar volume is given by Vm = RT /p = NAv V /N .
With NAv kB = R and the molar mass M = NAv m we obtain
( " 3 #)
5 2M kB T 2 RT
Sm = R + ln (7.30)
2 NAv h2 NAv p

7.3 Nuclear Spin Partition Function


7.3.1 High-Temperature Limit
In the absence of a magnetic field, all nuclear spin states are degenerate,1 except for the very tiny
splittings that arise from J couplings between the nuclear spins themselves. Even if we consider
the largest magnetic fields available, it is safe to assume that all nuclear spin states are equally
populated down to temperatures of at least 1.5 K and that the contribution of nuclear spins to
the total energy is negligibly small. Of course, NMR spectroscopy relies on the fact that these
states are not exactly equally populated, but in the context of statistical thermodynamics, the
contribution to internal energy and the population differences are negligible.
Hence, in this high-temperature limit all nuclear spin states are fully accessible and the
number of accessible states equals the total number of nuclear spin states. This gives
Y
Zns = (2Ii + 1) , (7.31)
i

where the Ii are the nuclear spin quantum numbers for nuclei in the molecule. Magnetic
equivalence leads to degeneracy of nuclear spin levels, but does not influence the total number of
nuclear spin states. Since the term u/T in Eq. (5.11) is negligible and zns = Zns N , we have

X
s = N kB ln (2Ii + 1) . (7.32)
i

1We neglect nuclear quadrupole coupling, which averages in gases.


70 Partition Functions of Gases

This contribution to entropy is not generally negligible. Still it is generally ignored in textbooks,
which usually does not cause problems, as the contribution is constant under most conditions
where experiments are conducted and does not change during chemical reactions.

7.3.2 Symmetry Requirements


Nuclear spin states have another, more subtle effect that may prevent separation of state spaces.
We consider this effect for H2 . In this molecule, the electron wavefunction arises from two
electrons, which are fermions, and must thus be antisymmetric with respect to exchange of the
two electrons. In quantum-chemical computations this is ensured by using a Slater determinant.
Likewise, the nuclear wavefunction must be antisymmetric with respect to exchange of the two
protons, which are also fermions. The spin part is antisymmetric for the singlet state with total
nuclear spin quantum number F = 0,
1
ns,S = (|i |i) , (7.33)
2
and symmetric for the triplet state with F = 1, as can be seen by the wavefunctions of each of
the three triplet substates:

ns,T+ = |i (7.34)
1
ns,T0 = (|i + |i) (7.35)
2
ns,T = |i . (7.36)

The translational, vibrational, and electron wavefunction are generally symmetric with respect to
the exchange of the two nuclei. The rotational wavefunction is symmetric for even rotational
quantum numbers J and antisymmetric for odd quantum numbers. Hence, to ensure that the
generalized Pauli principle holds and the total wavefunction is antisymmetric with respect to
exchange of indistinguishable nuclei, even J can only be combined with the antisymmetric
nuclear spin singlet state and odd J only with the symmetric triplet state. The partition functions
for these two cases must be considered separately. For H2 we have
2
X
Zpara = (2J + 1) eJ(J+1)~ /2IkB T , (7.37)
Jeven

where gJ = 2J + 1 is the degeneracy of the rotational states and I is the moment of inertia, and
2
X
Zortho = 3 (2J + 1) eJ(J+1)~ /2IkB T , (7.38)
Jodd

where gI = 3 is the degeneracy of the nuclear spin states.


For H2 the (J = 0, F = 0) state is called para-hydrogen and the (J = 1, F = 1) state
ortho-hydrogen. At ambient temperature, both the (J = 0, F = 0) state and the (J = 1, F = 1)
state are, approximately, fully populated and thus, the four nuclear spin substates described by
Eqs. (7.33-7.36) are equally populated. Statistics then dictates a para-hydrogen to ortho-hydrogen
ratio of 1:3 and no macroscopic spin polarization in a magnetic field. The splitting between the
two states is
J=1,F =1 J=0,F =0 ~2
= 2rot = 178.98 K , (7.39)
kB kB I
where we have introduced a characteristic rotational temperature analogous to the characteristic
vibrational temperature for the harmonic oscillator in Eq. (6.50). At temperatures well below this
7.4 Rotational Partition Function 71

energy splitting, para-hydrogen is strongly enriched with respect to ortho-hydrogen. Equilibration


in a reasonable time requires a catalyst. Still, no macroscopic spin polarization in a magnetic
field is observed, as the two nuclear spins are magnetically equivalent and aligned antiparallel.
If, however, para-hydrogen is reacted with a molecule featuring a multiple bond, magnetic
equivalence of the two hydrogen atoms can be removed and in that case enhanced nuclear spin
polarization is observable (para-hydrogen induced polarization, PHIP ). We note that for 2 H2 the
combination of nuclear spin states and rotational states to an allowed state reverses, as deuterons
are bosons.

7.4 Rotational Partition Function


7.4.1 Rigid Rotor Assumption and Rotamers
Separation of the rotational partition function from the partition functions of other degrees of
freedom does not only require consideration of nuclear spin states, but also the assumption that
the moment of inertia is the same for all rotational states. This is generally true if the molecule
behaves as a rigid rotor. For small molecules consisting of only a few atoms, this is often a good
approximation. Larger molecules feature internal rotations, where a group of atoms rotates with
respect to the rest of the molecule. In general, internal rotations are torsions about rotatable
bonds, which are often not freely rotatable. The torsion potential has several minima and these
minima are separated by energy barriers with heights that are larger, but not much larger than
kB T . If we denote the number of potential minima for the ith rotatable bond with nmin,i , the
total number of rotameric states, short rotamers is
Y
nrot = nmin,i . (7.40)
i

Each rotamer has its own moment of inertia and, hence, its own set of states with respect to
total rotation of the molecule. Because the energy scales of internal and total rotations are not
well separated and because in larger molecules some vibrational modes may also have energies
in this range, the partition function is not usually separable for large molecules. In such cases,
insight into statistical thermodynamics can be best obtained by MD simulations. In the following,
we consider small molecules that can be assumed to behave as a rigid rotor. We first consider
diatomic molecules, where it certainly applies on the level of precision that we seek here.
The energy levels of a rigid rotor of a diatomic molecule are quantized by the rotational
quantum number J and given by

h2
J = J(J + 1) = hcBJ(J + 1) , (7.41)
8 2 I
where

I = r2 (7.42)

is the moment of inertia with the reduced mass and

h
B= (7.43)
8 2 Ic
is the rotational constant. After introducing the characteristic rotational temperature,

h2 hcB
rot = = (7.44)
8 2 IkB kB
72 Partition Functions of Gases

we have

J = J(J + 1)kB rot . (7.45)

As already mentioned, each rotational level has a degeneracy gJ = 2J + 1. If all nuclei


in the molecules are distinguishable (magnetically not equivalent), there is no correlation with
nuclear spin states. In that case we have
X
Zrot = (2J + 1) eJ(J+1)rot /T . (7.46)

For sufficiently high temperatures and a sufficiently large moment of inertia, the density of states
is sufficiently large for replacing the sum by an integral,
Z
T
Zrot (2J + 1) eJ(J+1)rot /T dJ = . (7.47)
0 rot
Deviations between the partition functions computed by Eq. (7.46) and Eq. (7.47) are visualized
in Figure 7.1. As state functions depend on ln Z, the continuum approximation is good for
T /rot  1, which applies to all gases, except at low temperatures for those that contain
hydrogen. At ambient temperature it can be used in general, except for a further correction that
we need to make because of symmetry considerations.

15 A -0.5 B
ln(Zrot,approx/Zrot,exact)

-1
10
-1.5
Zrot

-2
5
-2.5

0 -3
0.1 0.3 1 3 10 0.1 0.3 1 3 10
T/Qrot T/Qrot

Figure 7.1: Continuum approximation for the rotational partition function. (A) Rotational partition
function obtained by the sum expression (7.46) (black line) and by the integral expression (7.47)
corresponding to the continuum approximation (red line). (B) Logarithm of the ratio between the rotational
partition functions obtained by the continuum approximation (7.47) and the exact sum formula (7.46).

7.4.2 Accessible States and Symmetry


Even if all nuclei are magnetically inequivalent and, hence, distinguishable, rotational states may
be not. For heteronuclear diatomic molecules they are, but for homonuclear diatomic molecules,
they are not. To see this, we consider a symmetrical linear molecule that rotates by 180 about
an axis perpendicular to the bond axis and centered in the middle of the bond. This rotation
produces a configuration that is indistinguishable from the original configuration. In other words,
the nuclear wavefunction is symmetric with respect to this rotation. For a homonuclear diatomic
molecule, half of the rotational states are symmetric and half are antisymmetric. For nuclei
that are bosons, such as 16 O in dioxygen, only the former states are allowed, for nuclei that are
fermions, only the latter are allowed. Hence, we need to divide the partition function, which
7.4 Rotational Partition Function 73

counts accessible states, by two. In this example, we have deliberately chosen a case with only
one nuclear spin state. If nuclear spin states with different symmetry exist, all rotational states
are accessible, but correlated to the nuclear spin states, as we have seen before for dihydrogen. In
the following we consider the situation with only one nuclear spin state or for a fixed nuclear spin
state.

Group Group Group Group


C1 , Ci , Cs , Cv 1 Dh 2 T , Td 12 Oh 24
Cn , Cnv , Cnh n Dn , Dnh , Dnd 2n Sn n/2 Ih 60

Table 7.1: Symmetry numbers corresponding to symmetry point groups [Her45; Iri98].

Although we still did not discuss other complications for multi-atomic molecules, we
generalize this concept of symmetry-accessible states by introducing a symmetry number . In
general, is the number of distinct orientations of a rigid molecule that are distinguished only
by interchange of identical atoms. For an NH3 molecule, rotation about the C3 axis by 120
generates one such orientation from the other. No other rotation axis exists. Hence, = 3 for
NH3 . In general, the symmetry number can be obtained from the molecules point group [Her45;
Iri98], as shown in Table 7.1.
With the symmetry number, the continuum approximation (7.47) becomes

T
Zrot = , (7.48)
rot
where we still assume that symmetry is sufficiently high for assigning a single characteristic
temperature.
We further find
1
ln Zrot = ln T + ln , (7.49)
rot
and with this
 
ln Zrot
2
Urot = NAv kB T (7.50)
T V

= RT 2 ln T (7.51)
T
= RT (linear) . (7.52)

On going from Eq. (7.50) to (7.51) we could drop the second term on the right-hand side of Eq.
(7.49), as this term does not depend on temperature. This is a general principle: Constant factors
in the partition function do not contribute to internal energy. The result can be generalized to
multi-atomic linear molecules that also have two rotational degrees of freedom. This result is
expected from the equipartition theorem, as each degree of freedom should contribute a term
kB T /2 to the molecular or a term RT /2 to the molar internal energy. However, if we refrain
from the continuum approximation in Eq. (7.47) and numerically evaluate Eq. (7.46) instead,
we find a lower contribution for temperatures lower than or comparable to rot . This is also a
general principle: Contributions of modes to internal energy and, by inference, to heat capacity,
are fully realized only at temperatures much higher than their characteristic temperature and are
negligible at temperatures much lower than their characteristic temperature.2
2The zero-point vibrational energy is an exception from this principle with respect to internal energy, but not heat
capacity.
74 Partition Functions of Gases

For the rotation contribution to molar heat capacity at constant volume of a linear molecule
we obtain

Crot,V = Urot = R . (7.53)
T
A non-linear multi-atomic molecule has, in general, three independent rotational moments of
inertia corresponding to three pairwise orthogonal directions a, b, c. With
h2
rot, = ( = a , b , c) (7.54)
8 2 I kB
one finds for the partition function
 1/2
T3
Zrot = . (7.55)
rot,a rot,b rot,c
For spherical-top molecules, all three moments of inertia are equal, Ia = Ib = Ic , and hence all
three characteristic temperatures are equal. For symmetric-top molecules, Ia = Ib 6= Ic .
The general equations for Urot and Crot,V at sufficiently high temperature T  rot are
d
Urot = RT (7.56)
2
d
Crot,V = R , (7.57)
2
where d = 1 for a free internal rotation (e.g., about a CC bond), d = 2 for linear, and d = 3 for
non-linear molecules. We note that the contribution of a free internal rotation needs to be added
to the contribution from total rotation of the molecule.
The expressions for the rotational contributions to molar entropy are a bit lengthy and do not
provide additional physical insight. They can be easily derived from the appropriate expressions
for the rotational partition function and internal energy using Eq. (5.11). We note, however,
that the contribution from the u/T term in the entropy expression is dR/2 and the contribution
from ln Z is positive. Hence, at temperatures where the continuum approximation is valid, the
rotational contribution to entropy is larger than dR/2.

7.5 Vibrational Partition Function


7.5.1 The Harmonic Oscillator Extended
Vibration in a diatomic molecule can be described by the 1D harmonic oscillator that we have
considered in Section 6.3.2. In a multi-atomic molecule the 3n 5 (linear) or 3n 6 (non-linear)
normal modes can be treated independently,
3n53n6 3n53n6
Y Y 1
Zvib = Zvib,i = . (7.58)
i=1 i=1
1 evib,i /T
Normal mode energies are no longer independent and the partition function is no longer
factorisable if anharmonicity of the vibration needs to be included, which is the case only at very
high temperatures. We ignore this and ask about the limiting behavior of Zvib for a diatomic
molecule or Zvib,i for an individual normal mode at high temperatures. In the denominator of
Eq. (6.51) we can make the approximation evib /T = 1 vib /T , if vib /T  1. We obtain
T
lim Zvib,i = . (7.59)
T vib,i
Vibrational temperatures for most normal modes are much higher than ambient temperature.
Hence, at 298 K we have often Zvib,i 1. Appreciable deviations are observed for vibrations
that involve heavy atoms, for instance Zvib = 1.556 at T = 300 K for I2 [Mac98].
7.6 Electronic Partition Function 75

7.5.2 Vibrational Contributions to U , CV , and S


The vibrational partition function for a system consisting of N diatomic molecules is
 N
N 1
zvib = Zvib = . (7.60)
1 evib /T
With N = NAv we obtain the vibrational contribution to the molar internal energy
 
ln zvib NA kB vib
Uvib = kB T 2 = /T (7.61)
T V e vib 1
Rvib
= /T . (7.62)
e vib 1
For multi-atomic molecules the contributions from the individual normal modes with characteristic
vibrational temperatures vib,i must be summed. Eq. (7.62) neglects the zero-point energy,
as we had defined the partition function for an energy scaled by the zero-point energy. On an
absolute energy scale, where U = 0 corresponds to the minimum of the Born-Oppenheimer
potential energy hypersurface, an additional term Uzp = NAv hi /2 needs to be added for each
normal mode, with i being the frequency of the normal mode.
The vibrational contribution to molar heat capacity at constant volume is

vib 2 evib /T
   
Uvib
Cvib,V = =R 2 , (7.63)
T V T evib /T 1
which is called the Einstein equation. With the Einstein function,
u2 eu
FE (u) = , (7.64)
(eu 1)2
it can be written as
 
vib
Cvib,V = RT FE . (7.65)
T
For computing the vibrational contribution to molar entropy we revert to the shifted energy
scale. This is required, as inclusion of the zero-point contribution to u would leave us with an
infinity. We find
" #
vib,i  
Svib,i = R  ln 1 evib,i /T . (7.66)
T evib /T 1

Again contributions from individual normal modes add up. For vib,i /T  1, which is the
usual case, both terms in the brackets are much smaller than unity, so that the contribution of any
individual normal mode to entropy is much smaller than R. Hence, at ambient temperature the
vibrational contribution to entropy is negligible compared to the rotational contribution unless
the molecule contains heavy nuclei.

7.6 Electronic Partition Function


Atoms and molecules can also store energy by populating excited electronic states. For the
hydrogen atom or any system that contains only a single electron, the energy levels can be given
in closed form, based on the Bohr model,
zq2 RE
el,n = , (7.67)
n2
76 Partition Functions of Gases

where n = 0 . . . is the principal quantum number, zq the nuclear charge, and RE the Rydberg
constant. However, this is an exception. For molecules and all other neutral atoms closed
expressions for the energy levels cannot be found.
In most cases, the problem can be reduced to considering either only the electronic ground
state with energy el,0 or to considering only excitation to the first excited state with energy
el,1 . If we use an energy shift to redefine el,0 = 0, we can define a characteristic electronic
temperature
el,1
el = . (7.68)
kB
Characteristic electronic temperatures are usually of the order of several thousand Kelvin. Hence,
in most cases, el  T applies, only the electronic ground state is accessible, and thus

Zel = gel,0 , (7.69)

where gel,0 is the degeneracy of the electronic ground state. We note that spatial degeneracy of
the electronic ground state cannot exist in non-linear molecules, according to the Jahn-Teller
theorem. However, a spatially non-degenerate ground state can still be spin-degenerate.
In molecules, total orbital angular momentum is usually quenched ( = 0, ground state).
In that case
{}
Zel = 2S + 1 , (7.70)

where S is the electron spin quantum number. For the singlet ground state of a closed-shell
{}
molecule (S = 0) we have Zel = 1, which means that the electronic contribution to the
partition function is negligible. The contribution to internal energy and heat capacity is generally
negligible for el  T . The electronic contribution to molar entropy,
{}
Sel = R ln (2S + 1) , (7.71)

is not negligible for open-shell molecules or atoms with S > 0. At high magnetic fields and low
temperatures, e.g. at T < 4.2 K and B0 = 3.5 T, where the high-temperature approximation for
electron spin states does no longer apply, the electronic partition function and corresponding
energy contribution are smaller than given in Eq. (7.70). For a doublet ground state (S = 1/2)
the problem can be solved with the treatment that we have given in Section 6.3.1. For S > 1/2
the spin substates of the electronic ground state are not strictly degenerate even at zero magnetic
field, but split by the zero-field splitting, which may exceed thermal energy in some cases. In that
case Eq. (7.70) does not apply and the electronic contribution to the partition function depends
on temperature. Accordingly, there is a contribution to internal energy and to heat capacity.
For a > 0 species with term symbol 2S+1 , each component is doubly degenerate
[SHM12]. For instance, for NO with a ground state ( = 1), both the 2 1/2 and the 2 3/2
state are doubly degenerate. As the 2 3/2 state is only 125 cm1 above the ground state, the
characteristic temperature for electronic excitation is el = 178 K. In this situation, Eq. (7.69)
does not apply at ambient temperature. The energy gap to the next excited state, on the other
hand, is very large. Thus, we have

Zel = gel,0 + gel,1 eel /T . (7.72)

This equation is fairly general, higher excited states rarely need to be considered. The electronic
contribution to the heat capacity of NO derived from Eq. (7.72) is in good agreement with
experimental data from Eucken and dOr [Mac98].
7.7 Equilibrium Constant for Gas Reactions 77

7.7 Equilibrium Constant for Gas Reactions


For clarity, we use an example reaction [WF12]

|A | A + |B | B |C | C + |D | D (7.73)

with adaptation to other reactions being straightforward. At equilibrium we must have

G = 0 , (7.74)

hence
X
i i = A A + B B + C C + D D = 0 , (7.75)
i

where the i are molar chemical potentials. To solve this problem, we do not need to explicitly
work with the grand canonical ensemble, as we can compute the i from the results that we have
already obtained for the canonical ensemble. According to one of Gibbs fundamental equations,
which we derived in the lecture course on phenomenological thermodynamics, we have
X
df = sdT pdV + i dni . (7.76)
i

Comparison of coefficients with the total differential of f (T, V, ni ) reveals that


 
f
i = , (7.77)
ni T,V,nj6=i

a result that we had also obtained in the lecture course on phenomenological thermodynamics.
Using Eq. (5.21), Eq. (5.18), and Stirlings formula, we obtain for the contribution fi of an
individual chemical species to Helmholtz free energy

fi = kB T ln zi (7.78)
1 Ni
= kB T ln Z (7.79)
Ni ! i
= kB T (Ni ln Zi Ni ln Ni + Ni ) (7.80)
Zi
= ni RT ln ni RT , (7.81)
ni NAv
where ni is the amount of substance (mol). Eq. (7.77) then gives
1 Zi
i = ni RT RT ln RT (7.82)
ni ni NAv
Zi
= RT ln (7.83)
ni NAv
Zi
= RT ln . (7.84)
Ni
Eq. (7.84) expresses the dependence of the chemical potential, a molar property, on the
molecular partition function. It may appear odd that this property depends on the absolute
number of molecules Ni , but exactly this introduces the contribution of mixing entropy that
counterbalances the differences in standard chemical potentials i . Because of our habit of
shifting energies by el,0 and by the zero-point vibration energies, we cannot directly apply Eq.
(7.84). We can avoid explicit dependence on the el,0,i and the zero-point vibrational energies by
78 Partition Functions of Gases

relying on Hess law and referencing energies of all molecules to the state where they are fully
dissociated into atoms. The energies i,diss for the dissociated states can be defined at 0 K. We
find

e(ij i,diss )/kB T


X
Zi,corr = (7.85)
j
X
=e i,diss /kB T
eij /kB T (7.86)
j

= Zi ei,diss /kB T , (7.87)

where index j runs over the states of molecule i.


With this correction we have
X Zi ei,diss /kB T
G = RT i ln . (7.88)
Ni
i

For our example reaction, the equilibrium condition is

A A + B B = C C D D , (7.89)

which gives

ZA eA,diss /kB T ZB eB,diss /kB T


RT A ln RT B ln
NA NB
ZC e C,diss /kB T ZD e D,diss /kB T
=RT C ln + RT D ln (7.90)
NC ND
and can be rearranged to
A
ZA eA A,diss /kB T ZBB eB B,diss /kB T ZCC eC C,diss /kB T ZD
D
eD D,diss /kB T
ln = ln
NAA NBB NCC NDD
(7.91)

and further rearranged to


| | | | | | | |
NC C ND D ZC C ZD D
| | | |
= | | | |
e(A A,diss +B B,diss +C C,diss +D D,diss )/kB T . (7.92)
NA A NB B ZA A ZB B

In Eq. (7.92) we can make the identifications


| | | |
NC C ND D
KN (V, T ) = | | | |
, (7.93)
NA A NB B

where KN (V, T ) is a volume-dependent equilibrium constant expressed with particle numbers,


and

U0 = NAv (A A,diss + B B,diss + C C,diss + D D,diss ) , (7.94)

where U0 is the molar reaction energy at 0 K. Hence, we have


| | | |
ZC C ZD D
KN (V, T ) = | | | |
eU0 /RT . (7.95)
ZA A ZB B
7.7 Equilibrium Constant for Gas Reactions 79

The dependence on volume arises from the dependence of the canonical partition functions on
volume.
i
By dividing all particle numbers by NAv and volume V i , we obtain the equilibrium constant
Kc (T ) in molar concentrations
| | | |
ZC C ZD D P
Kc (T ) = | | | |
(NAv V ) i i
eU0 /RT . (7.96)
ZA A ZB B
P
By dividing them by the total particle number N = i Ni to the power of i we obtain
| | | |
ZC C ZD D P
Kx (V, T ) = | | | |
N i i
eU0 /RT , (7.97)
ZA A ZB B

which coincides with the thermodynamical equilibrium constant K at the standard pressure p .
The most useful equilibrium constant for gas-phase reactions is obtained by inserting pi = ci RT
into Eq. (7.96)3:
| | | | Pi i
ZC C ZD D

RT
Kp (T ) = | | | |
eU0 /RT . (7.98)
ZA A ZB B NAv V

For each molecular species, the molecular partition function is a product of the contributions
from individual modes, Eq. (7.7), that we have discussed above. In the expression for equilibrium
constants, the nuclear-spin contribution cancels out since the number of nuclei and their spins
are the same on both sides of the reaction equation. Symmetry requirements on the nuclear
wavefunction are considered in the symmetry numbers i for the rotational partition function.
The electronic contribution often reduces to the degeneracy of the electronic ground state and in
the vibrational contribution, normal modes with vib,i > 5T can be neglected.

3There is a misprint in [WF12]


Thermodynamics of Mixing
Entropy of Binary Mixing
Energy of Binary Mixing
Entropic Elasticity
Ideal Chain Model
Random Walk
Conformational Entropy and Free Energy

8 Macromolecules

8.1 Thermodynamics of Mixing


The formalism introduced in Chapter 7 is suitable for small molecules in the gas phase, but does
not easily extend to condensed phases or to larger molecules with several rotameric states, let
alone to macromolecules, such as synthetic polymers, proteins, nucleic acids, and carbohydrates.
Nevertheless, statistical thermodynamics is an important theory for understanding such systems.
In this Chapter we introduce some of the concepts of statistical thermodynamics that do not
depend on explicit computation of the partition function. We start with the entropy of mixing
and, for simplicity, restrain the discussion to binary mixtures [RC03].

8.1.1 Entropy of Binary Mixing


We consider mixing of two species A with volume VA and B with volume VB and neglect volume
change, so that the total volume is VA + VB . The volume fractions of the two components in the
mixture are thus given by

VA
A = (8.1)
VA + VB
VB
B = = 1 A . (8.2)
VA + VB

To consider the statistics of the problem we use a lattice model.

Concept 8.1.1 Lattice model. A lattice model is a discrete representation of a system as


opposed to a continuum representation. A three-dimensional lattice model is a regular
arrangement of sites in Cartesian space, such as a crystal lattice is a regular arrangement of
atoms in Cartesian space. The state of the model is defined by the distribution of units of
matter, for instance molecules or the repeat units of a polymer (short: monomers), on the lattice
sites. In statistical thermodynamics, one particular arrangement of the units on the lattice is a
microstate. Energy of the microstate depends on interactions of units between lattice sites, in
the simplest case only between direct neighbor sites. By considering the statistical distribution
of microstates, thermodynamic state functions of the macrostate of the system can be obtained.
82 Macromolecules

In our example we assign the lattice site a volume v0 , which cannot be larger than the volume
required for one molecule of the smaller component in the mixture. The other component
may then also occupy a single site (similarly sized components) or several lattice sites. A
macromolecule with a large degree of polymerization consists of a large number of monomers
and will thus occupy a large number of lattice sites. The molecular volumes of the species are

vA = NA v0 (8.3)
vB = NB v0 , (8.4)

where NA and NB are the number of sites occupied by one molecule of species A and B,
respectively. We consider the three simple cases listed in Table 8.1. Regular solutions are
mixtures of two low molecular weight species with NA = NB = 1. Polymer solutions are
mixtures of one type of macromolecules (NA = N  1) with a solvent, whose molecular
volume defines the lattice site volume v0 (NB = 1). Polymer blends correspond to the general
case 1 6= NA 6= NB 6= 1. They are mixtures of two different species of macromolecules, so that
NA , NB  1.

NA NB
Regular solutions 1 1
Polymer solutions N 1
Polymer blends NA NB

Table 8.1: Number of lattice sites occupied per molecule in different types of mixtures [RC03].

The mixture occupies


VA + VB
n= (8.5)
v0
lattice sites, whereas component A occupies VA /v0 = nA of these sites. We consider a
microcanonical ensemble and can thus express entropy as

s = kB ln , (8.6)

where is the number of ways in which the molecules can be arranged on the lattice (number of
microstates). In a homogeneous mixture, a molecule or monomer of component A can occupy
any of the n lattice sites. Before mixing, it can occupy only one on the lattice sites in volume VA .
Hence, the entropy change for one molecule of species A is

SA = kB ln n kB ln A n (8.7)
n
= kB ln (8.8)
A n
= kB ln A . (8.9)

The total mixing entropy for both species is

smix = kB (nA ln A + nB ln B ) . (8.10)

We note the analogy with the expression that we had obtained in phenomenological thermody-
namics for an ideal mixture of ideal gases, were we had used the molar fraction xi instead of the
volume fraction i . For ideal gases, Vi ni and thus i = xi . Eq. (8.10) generalizes the result
to any ideal mixture in condensed phase. The mixture is ideal because we did not yet consider
energy of mixing and thus could get away with using a microcanonical ensemble.
8.1 Thermodynamics of Mixing 83

For discussion it is useful to convert the extensive quantity smix to the intensive entropy of
mixing per lattice site,
 
A B
S mix = kB ln A + ln B , (8.11)
NA NB

where we have used the number of molecules per species ni = ni /Ni and normalized by the
total number n of lattice sites.
For a regular solution with NA = NB = 1 we obtain the largest entropy of mixing at given
volume fractions of the components,

S mix = kB (A ln A + B ln B ) (regular solutions) . (8.12)

For a polymer solution with NA = N  1 and NB = 1 we have


 
A
S mix = kB ln A + B ln B (8.13)
N
kB B ln B , (8.14)

where the approximation by Eq. (8.14) holds for B  1/N , i.e. for solving a polymer and
even for any appreciable swelling of a high-molecular weight polymer by a solvent. For polymer
blends, Eq. (8.11) holds with NA , NB  1. Compared to formation of a regular solution or a
polymer solution, mixing entropy for a polymer blend is negligibly small, which qualitatively
explains the difficulty of producing such polymer blends. Nevertheless, the entropy of mixing is
always positive, and thus the Helmholtz free energy F mix = T S mix always negative, so
that an ideal mixture of two polymers should form spontaneously. To see what happens in real
mixtures, we have to consider the energetics of mixing.
Before doing so, we note the limitations of the simple lattice model. We have neglected
conformational entropy of the polymer, which will be discussed in Section 8.2.3. This amounts to
the assumption that conformational entropy does not change on mixing. For blends of polymers,
this is a very good assumption, whereas in polymers solutions there is often an excluded volume
that reduces conformational space. We have also neglected the small volume change that occurs
on mixing, most notably for regular solutions. For polymer solutions and blends this volume
change is very small.

8.1.2 Energy of Binary Mixing


To discuss the internal energy contribution to the free energy of mixing, we continue using the
simplified lattice model. In particular, we consider mixing at constant volume and we assume
that attractive or repulsive interactions between lattice sites are sufficiently small to not perturb
random distributions of solvent molecules and monomers on lattice sites. We also ignore that the
polymer chain is connected, as this would exclude random distribution of the monomers to the
lattice sites. Regular solution theory, as we consider it here, is a mean-field approach where the
interaction at a given lattice site is approximated by a mean interaction with the other lattice sites.
This neglects correlations. Although the model may appear crude (as many models in polymer
physics), it provides substantial insight and an expression that fits experimental data surprisingly
well (as is the case for many crude models in polymer physics).
We start by defining three pairwise interaction energies uAA , uAB , and uBB between adjacent
sites of the lattice. For random distribution, the probability that a molecule or monomer A
has a neighbor A is A and the probability that it has a neighbor B is 1 A . We neglect
boundary effects, as the ratio between the number of surface sites and inner sites is very small for
84 Macromolecules

a macroscopic system. The mean-field interaction energy per lattice site occupied by an A unit is
thus

UA = A uAA + (1 A ) uAB (8.15)

and the corresponding expression for a lattice site occupied by a B unit is

UB = A uAB + (1 A ) uBB . (8.16)

To continue, we need to specify the lattice, as the number of sites a adjacent to the site
under consideration depends on that. For a cubic lattice we would have a = 6. We keep a as a
parameter in the hope that we can eliminate it again at a later stage. If we compute a weighted
sum of the expressions (8.15) and (8.15) we double count each pairwise interaction, as we will
encounter it twice. Hence, total interaction energy of the mixture is
an
u= [A UA + (1 A ) UB ] , (8.17)
2
where we have used the probability A of encountering a site occupied by a unit A and (1 A )
of encountering a site occupied by a unit B. By inserting Eqs. (8.15) and (8.16) into Eq. (8.17)
and abbreviating A = , we obtain
an
u= { [uAA + (1 ) uAB ] + (1 ) [uAB + (1 ) uBB ]} (8.18)
2 h
an 2 i
= uAA + 2 (1 ) uAB + (1 )2 uBB . (8.19)
2
Before mixing the interaction energy per site in pure A is auAA /2 and in B auBB /2. Hence,
the total interaction energy before mixing is
an
u0 = [uAA + (1 ) uBB ] , (8.20)
2
so that we obtain for the energy change u = u u0 on mixing
an h 2 i
u = uAA + 2 (1 ) uAB + (1 )2 uBB uAA (1 ) uBB (8.21)
2
an  2
uAA + 2 (1 ) uAB + 1 2 + 2 1 + uBB
  
= (8.22)
2
an
= [ ( 1) uAA + 2 (1 ) uAB + ( 1) uBB ] (8.23)
2
an
= (1 ) (2uAB uAA uBB ) . (8.24)
2
We again normalize by the number n of lattice sites to arrive at the energy change per site on
mixing:
a
U mix = (1 ) (2uAB uAA uBB ) . (8.25)
2
For discussion we need an expression that characterizes the mixing energy per lattice site as
a function of composition and that can be easily combined with the mixing entropy to free
energy. The Flory interaction parameter,
a 2uAB uAA uBB
= , (8.26)
2 kB T
elegantly eliminates the number of adjacent lattice sites and provides just such an expression:

U mix = (1 ) kB T . (8.27)
8.2 Entropic Elasticity 85

Introducing such a parameter is an often-used trick when working with crude models. If the
parameter is determined experimentally, the expression may fit data quite well, because part of
the deviations of reality from the model can be absorbed by the parameter and its dependence on
state variables. We finally obtain the Flory-Huggins equation for the Helmholtz free energy of
mixing, F mix = U mix T S mix ,
 
1
F mix = kB T ln + ln (1 ) + (1 ) . (8.28)
NA NB

As the entropy contribution (first two terms in the brackets on the right-hand side of Eq.
(8.28)) to F mix is always negative, entropy always favors mixing. The sign of F mix depends
on the sign of the Flory parameter and the ratio between the energy and entropy. The Flory
parameter is negative and thus favors mixing, if 2uAB < uAA + uBB , i.e., if the interaction in
AB pairs is more attractive than the mean interaction in AA and and BB pairs. Such cases occur,
but are rare. In most cases, the Flory parameter is positive. Since the entropy terms are very
small for polymer blends, such blends tend to phase separate. In fact, high molecular weight
poly(styrene) with natural isotope abundance phase separates from deuterated poly(styrene).

8.2 Entropic Elasticity


8.2.1 Ideal Chain Model
Most polymer chains have rotatable bonds as well as bond angles along the polymer backbone that
differ from 180 . This leads to flexibility of the chain. Even if the rotations are not free, but give
rise to only nrot rotameric states per rotatable bond, the number of possible chain conformations
becomes vast. For Nrot rotatable bonds, the number of distinct conformations is nN rot
rot . The
simplest useful model for such a flexible chain is the freely jointed chain model [RC03]. Here we
assume bond vectors that all have the same length l = |~ri |, where ~ri is the bond vector of the ith
bond. If we further assume an angle ij between consecutive bond vectors, we can write the
scalar product of consecutive bond vectors as

~ri ~rj = l2 cos ij . (8.29)

This scalar product is of interest, as we can use it to compute the mean-square end-to-end distance
hR2 i of an ensemble of chains, which is the simplest parameter that characterizes the spatial
dimension of the chain. With the end-to-end distance vector of a chain with n bonds,
n
X
~n =
R ~ri , (8.30)
i=1

we have

~ n2 i
hR2 i = hR (8.31)
~n R
= hR ~ ni (8.32)
* n ! n +
X X
= ~ri ~rj (8.33)
i=1 j=1
n X
X n
= h~ri ~rj i . (8.34)
i=1 j=1
86 Macromolecules

By using Eq. (8.29) we find


n X
X n
hR2 i = l2 hcos ij i . (8.35)
i=1 j=1

In the freely jointed chain model we further assume that there are no correlations between
the directions of different bond vectors, hcos ij i = 0 for i 6= j. Then, the double sum in Eq.
(8.35) has only n non-zero terms for i = j with cos i i = 1. Hence,

hR2 i = nl2 . (8.36)

This again appears to be a crude model, but we shall now rescue it by redefining l. In an ideal
polymer chain we can at least assume that there is no interaction between monomers that are
separated by many other monomers,

lim hcos ij i = 0 . (8.37)


|ij|

Furthermore, for a given bond vector ~ri the sum over all correlations with other bond vectors
converges to some finite number that depends on i,
n
X
hcos ij i = C 0 (i) . (8.38)
j=1

Therefore, when including the correlations, Eq. (8.35) can still be simplified to
n
X
2
hR i = l 2
C 0 (i) = Cn nl2 , (8.39)
i=1

where Florys characteristic ratio Cn is the average value of C 0 (i) over all backbone bonds of
the chain.
In general, Cn depends on n, but for very long chains it converges to a value C . For
sufficiently long chains, we can thus approximate

hR2 i nC l2 , (8.40)

which has the same dependence on n and l as the crude model of the freely jointed chain, Eq.
(8.36). Hence, we can define an equivalent freely jointed chain with N Kuhn segments of length
b. From

hR2 i = N b2 nC l2 (8.41)

and the length of the maximally stretched equivalent chain, the contour length Rmax ,

Rmax = N b , (8.42)

we obtain
2
Rmax
N= (8.43)
C nl2
and the Kuhn length

hR2 i C nl2
b= = . (8.44)
Rmax Rmax
8.2 Entropic Elasticity 87

Typical values of C for synthetic polymers range from 4.6 for 1,4-poly(isoprene) to 9.5 for
atactic poly(styrene) with corresponding Kuhn lengths of 8.2 to 18 , respectively.
At this point we have found the mean-square end-to-end distance as a parameter of an
equilibrium macrostate. If we stretch the chain to a longer end-to-end distance, it is no longer at
equilibrium and must have larger free energy. Part of this increase in free energy must come
from a decrease in entropy that stretching induces by reducing the number of accessible chain
conformations. It turns out that this entropic contribution is the major part of the increase in
free energy, typically 90%. The tendency of polymer chains to contract after they have been
stretched is thus mainly an entropic effect. To quantify it, we need a probability distribution for
the end-to-end vectors and to that end, we introduce a concept that is widely used in natural
sciences.

8.2.2 Random Walk


The freely jointed chain model explicitly assumes that the direction of the next Kuhn segment
is uncorrelated to the directions of all previous Kuhn segments. Where the chain end will be
located after the next step that prolongs the chain by one segment depends only on the location
of the current chain end. The freely jointed chain thus has aspects of a Markov chain. Each
prolongation step is a random event and the trajectory of the chain in space a random walk.

Concept 8.2.1 Random walk. Many processes can be discretized into individual steps. What
happens in the next step may depend on only the current state or also on what happened in
earlier steps. If it depends only on the current state, the process is memoryless and fits the
definition of a Markov chain. A Markov chain where the events are analogous steps in some
parameter space can be modeled as a random walk. A random walk is a mathematically
formalized succession of random steps. A random walk on a lattice, where each step can only
lead from a lattice point to a directly neighboring lattice point is a particularly simple model.

We can use the concept of a random walk in combination with the concepts of statistical
thermodynamics in order to solve the problem of polymer chain stretching and contraction. The
problem is solved if we know the dependence of Helmholtz free energy on the length of the
end-to-end vector. This, in turn, requires that we know the entropy and thus the probability
distribution of the length of the end-to-end vector. This probability distribution is given by the
number
p of possible random walks (trajectories) that lead to a particular end-to-end distance
~ 2
R .
For simplicity we start with a simpler example in one dimension that we can later extend to
three dimensions. We consider the standard example in this field, a drunkard who has just left a
pub. We assume that, starting at the pub door, he makes random steps forth and back along the
road. What is the probability P (N, x) that after N steps he is at a distance of x steps up the road
from the pub door? The problem is equivalent to finding the number W (N, x) of trajectories of
length N that end up x steps from the pub door and dividing it by the total number of trajectories.
Any such trajectory consists of N+ steps up the road and N steps down the road, with the
final position being x = N+ N [RC03]. The number of such trajectories is, again, given by a
binomial distribution (see Section 3.1.6)
(N+ + N )! N!
W (N, x) = = , (8.45)
N+ !N ! [(N + x) /2]! [(N x) /2]!
whereas the total number of trajectories is 2N , as the drunkard has two possibilities at each step.
Hence,
1 N!
P (N, x) = N
, (8.46)
2 [(N + x) /2]! [(N x) /2]!
88 Macromolecules

leading to
   
N +x N x
ln P (N, x) = N ln 2 + ln(N !) ln ! ln !. (8.47)
2 2
The last two terms on the right-hand side can be rewritten as
    x/2  
N +x N X N
ln ! = ln !+ ln + s and (8.48)
2 2 2
s=1
    x/2  
N x N X N
ln ! = ln ! ln +1s , (8.49)
2 2 2
s=1

which leads to
  x/2  
N X N/2 + s
ln P (N, x) = N ln 2 + ln(N !) 2 ln ! ln . (8.50)
2 N/2 + 1 s
s=1

We now assume a long trajectory. In the range where x  N , which is realized in an


overwhelming fraction of all trajectories, the numerator and denominator logarithms in the last
term on the right-hand side of Eq. (8.50) can be approximated by series expansion, ln(1 + y) y
for |y|  1, which gives
   
N/2 + s 1 + 2s/N
ln = ln (8.51)
N/2 + 1 s 1 2s/N + 2/N
   
2s 2s
= ln 1 + ln 1 + 2/N (8.52)
N N
4s 2
. (8.53)
N N
Hence,
x/2   x/2  
X N/2 + s X 4s 2
ln = (8.54)
N/2 + 1 s N N
s=1 s=1
x/2 x/2
4 X 2 X
= s 1 (8.55)
N N
s=1 s=1
4 (x/2)(x/2 + 1) x
= (8.56)
N 2 N
x2
= . (8.57)
2N
Inserting Eq. (8.57) into Eq. (8.50) provides,
x2
 
1 N!
P (N, x) N exp , (8.58)
2 (N/2)!(N/2)! 2N
where we recognize, in the last factor on the right-hand side, the approximation of the binomial
distribution by a Gaussian distribution that we discussed in Section 3.1.6. Using the improved
formula of Stirling, Eq. (3.31), for expressing the factorials, we have
r
1 N! 1 2N N N exp(N ) 2
= N  2 = , (8.59)
2N (N/2)!(N/2)! 2 N
N (N/2)N/2 exp(N/2)
8.2 Entropic Elasticity 89

which leads to the exceedingly simple result:


r
x2
 
2
P (N, x) = exp . (8.60)
N 2N
The drunkard, if given enough time and not falling into sleep, perfectly simulates a Gaussian
distribution.
We may even further simplify this result by asking about the mean square displacement hx2 i,
which is given by
Z r Z
x2
 
2 2 2 2
hx i = x P (N, x)dx = x exp dx = N . (8.61)
N 2N
Before we go on, we need to fix a problem that occurs when we interpret the discrete probabilities
computed at this point as a continuous probability density distribution of x. In the discrete case,
W (N, x) can be non-zero only for either even or odd x, depending on whether N is even or
odd. Thus, to arrive at the proper probability distribution we need to divide by 2. Hence, we can
express the probability density distribution for a one-dimensional random walk as
x2
 
1
1d (x) = p exp . (8.62)
2hx2 i 2hx2 i

This result does no longer depend on step size, not even implicitly, because we have removed the
dependence on step number N . Therefore, it can be generalized to three dimensions. Since the
random walks along the three pairwise orthogonal directions in Cartesian space are independent
of each other, we have

3d (x, y, z)dxdydz = 1d (x)dx 1d (y)dy 1d (z)dz . (8.63)

At this point we relate the result to the conformational ensemble of an ideal polymer chain,
using the Kuhn model discussed in D Section
E 8.2.1. We pose the question of the distribution of
~ 2
mean-square end-to-end distances R with the Cartesian components of the end-to-end vector
~ being x = Rx , y = Ry , and z = Rz . According to Eq. (8.41), we have
R
D E



~ 2 = Rx2 + Ry2 + Rz2
R (8.64)
= N b2 . (8.65)

For symmetry reasons we have,



2
2
2 N b2
Rx = Ry = Rz = , (8.66)
3
leading to
r
3Rx2
 
3
1d (N, x) = exp (8.67)
2N b2 2N b2
and analogous expressions for 1d (y) and 1d (z). We have reintroduced parameter N , which
is now the number of Kuhn segments. Yet, by discussing a continuous probability density
distribution, we have removed dependence on a lattice model. This is necessary since the steps
along dimensions x, y, and z differ for each Kuhn segment. By using Eq. (8.64), we find
3/2 !

3 3R~2
~ =
3d (N, R) exp . (8.68)
2N b2 2N b2
90 Macromolecules

0.01 1

Cumulative probability
0.008 A 0.8 B
r3d(N,R) [a.u.]

0.006 0.6

0.004 0.4

0.2
0.002

0
0
0 0.5 1 1.5 2 2.5 3 0 0.5 1 1.5 2 2.5 3
R/(bN ) R/(bN )

Figure 8.1: End-to-end distance distribution of an ideal


chain. (A) Probability density distribution
3d (N, R) of normalized en-to-end distances R/(b N ). (B) Cumulative probability distribution
indicating the probability to find a chain with end-to-end-distance smaller than R/(b B). Slightly
more
than 60% of all chains (red dashed horizontal line) have a normalized end-to-end distance R/(b N ) 1.

The probability density attains a maximum at zero end-to-end vector.


Finally, we can pose the following question: If we let all chains of the ensemble start at the
same point, how are the chain ends distributed in space? This is best pictured in a spherical
coordinate system. Symmetry dictates that the distribution is uniform with respect to polar angles
and . The polar coordinate R is equivalent to the end-to-end distance of the chain. To find
the probability distribution for this end-to-end distance we need to include area 4R2 of the
spherical shells. Hence,
3/2
3R2
  
3
3d (N, R) 4R2 dR = 4 exp R2 dR . (8.69)
2N b2 2N b2

Because of this scaling with the volume of an infinitesimally thin spherical shell, the probability
density distribution (Figure 8.1A) for the end-to-end distance does not peak
at zero distance. As
seen in Figure 8.1B it is very unlikely to encounter a chain with R > 2b N . Since the contour
length is Rmax = N b, we can conclude that at equilibrium almost all chains have end-to-end
distances shorter than 2Rmax / N .
We need to discuss validity of the result, because in approximating the discrete binomial
distribution by a continuous Gaussian probability distribution we had made the assumption
x  N . Within the ideal chain model, this assumption corresponds to an end-to-end distance
that is much shorter than the contour length N b. If R approaches N b, the Gaussian distribution
overestimates true probability density. In fact, the Gaussian distribution predicts a small, but
finite probability for the chain to be longer than its contour length, which is unphysical. The
model can be refined to include cases of such strong stretching of the chain [RC03]. For our
qualitative discussion of entropic elasticity not too far from equilibrium, we can be content with
Eq. (8.68).

8.2.3 Conformational Entropy and Free Energy


We may now ask the question of the dependence of free energy on chain extension R. ~ With the
definition of Boltzmann entropy, Eq. (5.3), and the usual identification k = kB we have

~ = kB ln (N, R)
S(N, R) ~ . (8.70)
8.2 Entropic Elasticity 91

The probability density distribution in Eq. (8.68) is related to the statistical weight by

~
~ = R (N, R) ,
3d (N, R) (8.71)
~ R
(N, R)d ~

because 3d is the fraction of all conformations that have an end-to-end vector in the infinitesimally
small interval between R ~ and R~ + dR.~ Hence,1
Z 
~ ~
S(N, R) = kB ln 3d (N, R) + kB ln ~
(N, R)dR ~ (8.72)

3 ~2
R 3

3
 Z 
= kB 2 + kB ln + kB ln ~ ~
(N, R)dR . (8.73)
2 Nb 2 2N b2

The last two terms do not depend on R ~ and thus constitute an entropy contribution S(N, 0) that
is the same for all end-to-end distances, but depends on the number of monomers N ,

~2
~ = 3 kB R + S(N, 0) .
S(N, R) (8.74)
2 N b2
Since by definition the Kuhn segments of an ideal chain do not interact with each other,
~ The Helmholtz free energy F (N, R)
the internal energy is independent of R. ~ = U (N, R)~
~
T S(N, R) can thus be written as

~2
~ = 3 kB T R + F (N, 0) .
F (N, R) (8.75)
2 N b2
It follows that the free energy of an individual chain attains a minimum at zero end-to-end vector,
in agreement with our conclusion in Section 8.2.2 that the probability density is maximal for
a zero end-to-end vector. At longer end-to-end vectors, chain entropy decreases quadratically
with vector length. Hence, the chain can be considered as an entropic spring. Elongation of
the spring corresponds to separating the chain ends by a distance R  N b. The force required
for this elongation is the derivative of Helmholtz free energy with respect to distance. For one
dimension, we obtain
 
F N, R ~
3kB T
fx = = Rx . (8.76)
Rx N b2
~
For the three-dimensional case, the force is a vector that is linear in R,
3kB T ~
f~ = R, (8.77)
N b2
i.e., the entropic spring satisfies Hookes law. The entropic spring constant is 3kB T /(N b2 ).
Polymers are thus the easier to stretch the larger their degree of polymerization (proportional
to N ), the longer the Kuhn segment b and the lower the temperature T . In particular, the
temperature dependence is counterintuitive. A polymer chain under strain will contract if
temperature is raised, since the entropic contribution to Helmholtz free energy, which counteracts
the strain, then increases.

1This separation of the terms is mathematically somewhat awkward, since in the last two terms the argument of
the logarithm has a unit. However, if the two terms are combined the logarithm of the unit cancels.
Bibliography

Books
[AP13] P. W. Atkins and J. de Paula. Physikalische Chemie. 5th Ed. Weinheim: WILEY-VCH,
2013 (cited on page 8).
[ER06] T. Engel and P. Reid. Physikalische Chemie. 1st Ed. Mnchen: Pearson Studium,
2006 (cited on page 8).
[Her45] G. Herzberg. Molecular Spectra and Molecular Structure: II. Infrared and Raman
Spectra of Polyatomic Molecules. New York: van Nostrand, 1945 (cited on page 73).
[Mac98] A. Maczek. Statistical Thermodynamics. First. Oxford: Oxford University Press,
1998 (cited on pages 8, 66, 68, 74, 76).
[Pen70] O. Penrose. Foundations of Statistical Mechanics. Oxford: Pergamon Press, 1970
(cited on pages 8, 12, 15, 16).
[Pre+97] W. H. Press et al. Numerical Recipes in C. Second. Cambridge: Cambridge University
Press, 1997 (cited on page 25).
[RC03] M. Rubinstein and R. H. Colby. Polymer Physics. First. Oxford: Oxford University
Press, 2003 (cited on pages 81, 82, 85, 87, 90).
[SHM12] M. Schfer, U. Hollenstein, and F. Merkt. Advance Physical Chemistry. Statistical
Thermodynamics. Autumn Semester 2012. Zrich: ETH Zrich, 2012 (cited on
page 76).
[Sch06] F. Schwabl. Statistische Mechanik. 3rd Ed. Berlin: Springer, 2006 (cited on pages 8,
14, 49, 53).
[SJ01] A. Schweiger and G. Jeschke. Principles of Pulse Electron Paramagnetic Resonance.
Oxford: Oxford University Press, 2001 (cited on page 52).
[Swe12] R. H. Swendsen. An Introduction to Statistical Mechanics and Thermodynamics. 1st
Ed. Oxford: Oxford University Press, 2012 (cited on pages 8, 18, 20, 22, 24, 25, 41,
52, 53).
[WF12] G. Wedler and H.-J. Freund. Lehrbuch der Physikalischen Chemie. 6th Ed. Weinheim:
WILEY-VCH, 2012 (cited on pages 8, 33, 37, 43, 44, 5457, 61, 62, 67, 77, 79).
94 Macromolecules

Articles
[Coh11] D. Cohen. Lecture Notes in Statistical Mechanics and Mesoscopics. In: arXiv:1107.0568
(2011), pages 1121 (cited on page 8).
[FR03] S. R. D. French and D. P. Rickles. Understanding Permutation Symmetry. In:
arXiv:quant-ph/0301020 (2003), pages 1 (cited on page 54).
[Kub63] R. Kubo. In: J. Math. Phys. 4 (1963), pages 174183 (cited on page 49).
[Tan06] Y. Tanimura. In: J. Phys. Soc. Jap. 75 (2006), page 082001 (cited on page 49).
[VF74] A. J. Vega and D. Fiat. In: J. Chem. Phys. 60 (1974), pages 579583 (cited on
page 49).

Web Pages
[Iri98] K. K. Irikura. Essential Statistical Thermodynamics. http://cccbdb.nist.gov/
thermo.asp. [Online; accessed 4-August-2015]. 1998 (cited on pages 9, 73).
[UNM12] UNM. Laminar flow. https://www.youtube.com/watch?v=_dbnH- BBSNo.
[Online; accessed 28-July-2015]. 2012 (cited on page 47).
Index

A D

accessible states . . . . . . . . . . . . . . . . . . . . . . . 68 degeneracy . . . . . . . . . . . . . . . . . . . . . . . . 58, 76


degree of . . . . . . . . . . . . . . . . . . . . . . . . . 58
density matrix
equilibrium . . . . . . . . . . . . . . . . . . . . . . . 53
B quantum mechanical . . . . . . . . . . . . . . . 51
density of states . . . . . . . . . . . . . . . . . 55, 62, 67
Bayes theorem . . . . . . . . . . . . . . . . . . . . . . . . 18 density operator . . . . . . . . . . . . . . . . . . . . . . . 51
binomial distribution . . . . . . . . . . . . . . . . . . . 22 Dirac delta function . . . . . . . . . . . . . . . . . . . . 27
Gaussian approximation . . . . . . . . . . . . 23 distinguishable variables . . . . . . . . . . . . . . . . 19
Boltzmann distribution . . . . . . . . . . 33, 36, 53 dynamical space . . . . . . . . . . . . . . . . . . . . . . . 11
Born-Oppenheimer approximation . . . . . . . 66 dynamical state . . . . . . . . . . . . . . . . . . . . . . . . 11
Bose-Einstein statistics . . . . . . . . . . . . . . . . . 54
boson . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
E

Einstein equation . . . . . . . . . . . . . . . . . . . . . . 75
C ensemble . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
canonical . . . . . . . . . . . . . . . . . . . . . . . . . 33
canonical ensemble . . . . . . . . . . . . . . . . . . . . 33
grand canonical . . . . . . . . . . . . . . . . . . . 39
central limit theorem . . . . . . . . . . . . . . . . . . . 23 microcanonical . . . . . . . . . . . . . . . . . . . . 32
central paradox . . . . . . . . . . . . . . . . . . . . . . . . 47 entropy . . . . . . . . . . . . . . . . . . . . . . . . 15, 41, 45
characteristic temperature Boltzmann . . . . . . . . . . . . . . . . . . . . . . . . 43
electronic . . . . . . . . . . . . . . . . . . . . . . . . . 76 Clausius . . . . . . . . . . . . . . . . . . . . . . . . . . 42
rotational . . . . . . . . . . . . . . . . . . . . . . . . . 70 Gibbs . . . . . . . . . . . . . . . . . . . . . . . . . 43, 49
vibrational . . . . . . . . . . . . . . . . . . . . . . . . 60 mixing . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
coherence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 Shannon . . . . . . . . . . . . . . . . . . . . . . . . . . 50
configuration . . . . . . . . . . . . . . . . . . . . . . . . . . 34 von Neumann . . . . . . . . . . . . . . . . . . . . . 49
contour length . . . . . . . . . . . . . . . . . . . . . . . . . 86 equipartition theorem . . . . . . . . . . . . . . . 37, 62
correlation function . . . . . . . . . . . . . . . . . . . . 21 ergodicity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
96 INDEX

Liouville-von-Neumann equation . . . . . . . . 15
Loschmidt paradox . . . . . . . . . . . . . . . . . . . . . 47
F

Fermi-Dirac statistics . . . . . . . . . . . . . . . . . . . 56 M
fermion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
Flory characteristic ratio . . . . . . . . . . . . . . . . 86 macrostate . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
Flory interaction parameter . . . . . . . . . . . . . 84 Markov chain . . . . . . . . . . . . . . . . . . . . . . 13, 87
Flory-Huggins equation . . . . . . . . . . . . . . . . 85 Markovian postulate . . . . . . . . . . . . . . . . . . . . 13
freely jointed chain . . . . . . . . . . . . . . . . . . . . . 85 Maxwell-Boltzmann statistics . . . . . . . . . . . 56
mean field approach . . . . . . . . . . . . . . . . 61, 83
mean value . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
G microcanonical ensemble . . . . . . . . . . . . . . . 32
microstate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
Gibbs paradox . . . . . . . . . . . . . . . . . . . . . . . . . 46 mixed state . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
grand canonical ensemble . . . . . . . . . . . . . . 39 mixing entropy . . . . . . . . . . . . . . . . . . . . . . . . 81
ground state . . . . . . . . . . . . . . . . . . . . . . . . . . . 53 moment analysis . . . . . . . . . . . . . . . . . . . . . . . 25
moment of inertia . . . . . . . . . . . . . . . . . . . . . . 71

H
N
Hamiltonian equations of motion . . . . . . . . 13
Heisenbergs uncertainty principle . . . . . . . 60 Newtonian equations of motion . . . . . . . . . . 11
Hess law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78

O
I
observational state . . . . . . . . . . . . . . . . . . . . . 12
ideal chain . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
independent events . . . . . . . . . . . . . . . . . . . . . 21
independent variables . . . . . . . . . . . . . . . . . . 19 P

para-hydrogen induced polarization . . . . . . 71


partition function
J canonical . . . . . . . . . . . . . . . . . . . . . . 36, 46
grand canonical . . . . . . . . . . . . . . . . . . . 39
Jahn-Teller theorem . . . . . . . . . . . . . . . . . . . . 76
molecular . . . . . . . . . . . . . . . . . . . . . . . . . 45
quantum mechanical . . . . . . . . . . . . . . . 53
K system . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
Pauli exclusion principle . . . . . . . . . . . . . . . . 54
Kronecker delta . . . . . . . . . . . . . . . . . . . . . . . . 20 Pauli principle . . . . . . . . . . . . . . . . . . . . . . . . . 70
Kuhn length . . . . . . . . . . . . . . . . . . . . . . . . . . . 86 Penrose postulates . . . . . . . . . . . . . . . . . . . . . 15
phase space . . . . . . . . . . . . . . . . . . . . . . . . 13, 13
Poisson brackets . . . . . . . . . . . . . . . . . . . . . . . 14
L population
quantum mechanical . . . . . . . . . . . . . . . 52
Lagrange multipliers . . . . . . . . . . . . . . . . . . . 35 probability . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
lattice model . . . . . . . . . . . . . . . . . . . . . . . . . . 81 conditional . . . . . . . . . . . . . . . . . . . . 18, 27
Liouville equation . . . . . . . . . . . . . . . . . . . . . 14 density . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
stochastic . . . . . . . . . . . . . . . . . . . . . . . . . 49 joint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
Liouville theorem . . . . . . . . . . . . . . . . . . . . . . 14 marginal. . . . . . . . . . . . . . . . . . . . . . .18, 26
INDEX 97

probability distribution . . . . . . . . . . . . . . . . . 20
moments . . . . . . . . . . . . . . . . . . . . . . . . . . 21
pure state . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 V

variance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
Q
W
quantum number
rotational . . . . . . . . . . . . . . . . . . . . . . 70, 71 Wiederkehreinwand . . . . . . . . . . . . . . . . . . . . 48
translational . . . . . . . . . . . . . . . . . . . . . . . 67
vibrational . . . . . . . . . . . . . . . . . . . . . . . . 59
Z

Zermelo pardox . . . . . . . . . . . . . . . . . . . . . . . . 48
R

random number
continuous . . . . . . . . . . . . . . . . . . . . . . . . 25
discrete . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
independent . . . . . . . . . . . . . . . . . . . . . . . 27
random walk . . . . . . . . . . . . . . . . . . . . . . . . . . 87
residual zero-point vibration . . . . . . . . . . . . 60
rotamers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71

Sackur-Tetrode equation . . . . . . . . . . . . . . . . 69
standard deviation . . . . . . . . . . . . . . . . . . . . . 20
state space . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
statistical weight . . . . . . . . . . . . . . . . . . . . . . . 32
statistics
Bose-Einstein . . . . . . . . . . . . . . . . . . . . . 54
Fermi-Dirac . . . . . . . . . . . . . . . . . . . . . . . 56
Maxwell-Boltzmann . . . . . . . . . . . . . . . 56
Stirlings formula . . . . . . . . . . . . . . . . . . . . . . 24
Gospers approximation . . . . . . . . . . . . 24
stochastic Liouville equation . . . . . . . . . . . . 49
symmetry number . . . . . . . . . . . . . . . . . . . . . 73

trajectory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

Umkehreinwand . . . . . . . . . . . . . . . . . . . . . . . 47

You might also like