Professional Documents
Culture Documents
Entropy Revisited, Gorilla and All
Entropy Revisited, Gorilla and All
men tioned once in the article, but in a context that does not recognize his seminal role in laying an axiomatic
foundation for thermodynamics (beginning more than 90 years ago!). The mathematical complexities of his formulation
of the second law not only obscured the physical simplicity of his idea, but for some time they were a significant
impediment to the use of his methods by physicists. Hans Buch dahl’s exposition of the Carathéodory formulation
2
played an extremely important role in popularizing the axiomatic development begun by Carathéodory, so that by the
late 1960s, there were several undergradu ate textbooks incorporating his approach. Although the authors do give
3,4
Buchdahl credit, there is even more reason to credit Carathéodory: He started the whole movement!
References
1. C. Carathéodory, Math. Ann. 67, 355 (1909); Sitzungsber. K. Preuss. Akad. Wiss. 39 (1925).
2. H. A. Buchdahl, Amer. J. Phys. 17, 41, 44, 212 (1949).
3. H. A. Buchdahl, The Concepts of Classi cal Thermodynamics, Cambridge U. Press, London (1966).
4. C. J. Adkins, Equilibrium Thermody namics, McGraw-Hill, London (1968). DAVID SIMINOVITCH
(siminovitch@hg.uleth.ca)
The University of Lethbridge
Lethbridge, Alberta, Canada
difficulties. But “normal” thermodynamics rules these possibilities out (usually implicitly). Regarding the symbol for
adiabatic accessibility, appropriately stressed on page 34 of the Lieb and Yngvason article, some credit might be given
to Constantin Carathéodory for an early use of this concept. His work need no longer be avoided on grounds of
mathematical complica tion, since Carathéodory’s principle can now be related to Kelvin’s princi ple, mentioned on
page 33 of the article, rather simply. Proceedings of the first international conference on thermodynamics provide
4 5
further citations.
To the (perhaps rhetorical) ques tion raised on page 32 of the article as to whether statistical mechanics
is essential to the second law, the answer is presumably “No,” since the main work of Sadi Carnot and Rudolph
Clausius preceded that of Willard Gibbs.
References
1. P. T. Landsberg, R. B. Mann, Class. Quantum Grav. 10, 2373 (1993). 2. R. D. Sorkin, R. M. Wald, Z. Z. Jiu,
Gen. Rel. Grav. 13, 1127 (1981). See also P. T. Landsberg, R. P. Woodward, J. Stat. Phys. 73, 361 (1993).
3. P. T. Landsberg, in The Study of Time III, J. T. Fraser, N. Lawrence, D. Park, eds., Springer, New York (1978), p. 118. P. T.
Landsberg, Seeking Ulti mates, Institute of Physics, Bristol, UK (2000), p. 121.
4. P. T. Landsberg, Nature 201, 485 (1964). 5. P. T. Landsberg, Pure Appl. Chem. 22, 215 (1970). Also published as a sepa rate
volume, Proceedings of the Interna tional Conference on Thermodynamics, Butterworths, London (1970).
PETER T. LANDSBERG
(ptl@maths.soton.ac.uk)
University of Southampton
Southampton, UK
thermodynamics makes entropy a necessary variable for the understanding of thermal, chemical, mechanical, and
electrical processes. It is interesting that the authors cred it Peter T. Landsberg for beginning the movement to use the
second law as a pillar of physics in its own right, but no interaction between Callen and Landsberg is mentioned in the
article.
To show the simplicity and beauty of Callen’s approach, a brief summary of his ideas follows. He first points out
that conservation of energy is not sufficient to explain many physical phenomena: A body at a uniform tem perature
does not spontaneously develop a gradient nor does a homo geneous chemical system separate
into components. An extensive vari able, the entropy S, is needed to explain why. Callen states that Entropy depends
on the internal energy U, volume t, number of con stituents N, and charge q of a sys tem through a relation of the
general form S = S(U, t, N, q), which may be inverted, in principle, to give U = U(S, t, N, q).
Entropy is defined only for equi librium states.
The entropy of a system for which all differences in temperature, pres sure, concentration, and so forth are allowed
to equalize will go to a maximum.
Differentiating the expression for U then leads to the absolute temper ature, the negative of the pressure, and the
chemical and electrostatic potentials. With those definitions in hand, the first law of thermodynam ics follows, and
includes thermal, mechanical, chemical, and electro static energy. It is then a simple matter to show that conservation
of energy combined with the entropy maximizing postulate leads to well known equilibrium conditions, such as
uniformity of temperature, chemi cal potential, or electrostatic poten tial. If N and q are constant, then dU = T dS – P dt,
which we may rewrite as dU = dQ – dW, where +dQ is the amount of heat that enters a body and +dW is the amount of
work done by the body. This is the primi tive form of the first law, and the relation dS = dQ/T can be applied to a
Carnot cycle to verify that all changes in an isolated system lead to an increase in the entropy. As the entropy of a 2
system
increases, so does a measure of its organization known as the disorder W. The connection between entropy and
disorder can be expressed as S = k log W e
where k is Boltzmann’s constant. An ingenious single combinatorial expression by Cowan, plus the use of
3
distinguishability and the exclu sion principle, shows that maximiz ing S with Lagrange multipliers leads to the
appropriate statistical distribution: Maxwell–Boltzmann, Fermi–Dirac, or Bose–Einstein. I agree with the authors, then,
that the Gibbs–Boltzmann approach is clearly the wrong direction to go, but it also follows that entropy can be
described and explained without all the complex mathematics.
References
1. H. B. Callen, Thermodynamics, Wiley, New York (1960).
2. G. Joos, Theoretical Physics, G. E.
Stechert, New York (1934).
3. R. D. Cowan, Am. J. Phys. 25, 463 (1957).
ALLEN NUSSBAUM
(nussbaum@ece.umn.edu)
are happy that our article attract ed the interest of so many letter writers and we appreciate their com ments. Yes, we
would have wished to include more references, but space limitations gave us little room for scholarship. However, our
long paper discusses the history and relationships of the various formula tions of the second law; we hoped that readers
1
would consult that paper. The recent paper of Jos Uffink is also valuable.
2
As mentioned by Siminovitch and Landsberg, and by us in our article, Carathéodory is one of the founding fathers, and
it is true that he was the first to emphasize the idea of a relation among states of macroscopic systems based on
adiabatic accessi bility. We regarded his ideas, which date to 1909, as well known, but per haps the connection could
have been clearer. However, although his work motivated Landsberg and the others
we mentioned, his conceptual frame work is different in important respects from the circle of ideas that led to the work
of Giles and then to our work.
Our work is logically divided into two parts. The first shows that entropy comes out of little more than the list of
pairs of states X and Y such that one can go from X to Y without doing more to the surround ings than moving a weight.
It is this fact that speaks for our approach (and Giles’s book). Quasistatic paths in state space are not needed; calcu lus
and the sophisticated mathemat ics of differential forms are also unnecessary in this part. It is not even necessary to
parametrize equi librium states by coordinates such as energy and volume. Nevertheless, entropy emerges together with
a spe cific formula that determines the entropy function uniquely, except for the choice of units. Carathéodory, on the
other hand, makes essential use of coordinates and differentials in state space, and entropy and temper ature appear to
be more delicate con structs than they naturally are. Likewise, we do not have to intro duce extraneous physical
principles and heat engines, as in the older approach of Carnot and successors.
What we must introduce in the first part is the notion of adiabatic accessibility, and since we do not have the help of
coordinates, we must do this empirically and without mentioning heat and temperature. This is where the gorilla enters
as a metaphor for the rest of the universe and its possible action on the system under discussion. Unlike Tourin, we do
not find this confusing. Rather, we consider it to be an essential clar ification of the kinds of processes with which the
second law deals.
Many formulations of the second law have been made, and choosing among them is largely a question of taste. We
submit, however, that our assumptions are easier to under stand than those of Kelvin, Planck, Clausius, and
Carathéodory. Callen’s approach, mentioned by Nussbaum, is excellent (as is the work of Tisza and ultimately Gibbs,
on which he relies), but Callen starts where we leave off. In his approach one postu lates the existence and basic proper
ties of entropy and then runs with the ball. We seek to answer the prior question: Where does entropy—and its
properties, especially its temporal increase—come from? Our answer is that it comes from a relation among
macroscopic equilibrium states
whose simple properties are, except for one, so obvious that they are taken for granted and rarely men tioned
explicitly.
The one nonobvious property of equilibrium states is the “compari son hypothesis,” which states that, given any
two states X and Y of a
system, either X is adiabatically accessible from Y or Y is adiabatical ly accessible from X. The second half of our work,
which goes beyond Giles, Landsberg, and the others we mentioned, turns this hypothesis into a fact, with the aid of
some rea sonable assumptions. Here, for the first time, we describe states by means of energy and volume, in the usual
way, and make some contact with Carathéodory’s approach. We introduce an important assumption that is similar to,
but significantly narrower than, Carathéodory’s main one: For every state X there is anoth er state, Y, somewhere, and
not nec essarily nearby, that is adiabatically accessible from X, but not the other way around. Otherwise, our approach is
mathematically and physically dif ferent from Carathéodory’s. It is not necessary to derive the entropy func tion afresh,
because that was done in the first part. We simply have to derive the simpler, nonquantita tive fact that any two states
are comparable.
The uniqueness of entropy implies that it can be measured conveniently without using the original formula that
established its existence. This answers the objections of Crowell and Tourin; you can find the differ ence of entropy
between two states in the old-fashioned way by measur ing specific heats, compressibilities, and so on. The uniqueness
guaran tees that all experiments will answer the question in the same way. Once we have the existence and unique ness
of entropy, all the techniques of the mechanical and chemical engi neers mentioned by Tourin are at our disposal.
We disagree with Crowell’s com ment that one can find the entropy of a macroscopic system by cutting it into tiny
subsystems. If, by this, he means that reduction to submeso scopic sizes will simplify thermody namics, we have to
demur. Clearly, additivity and scaling do not hold down to the atomic level because of surface effects. Even if, at some
suf ficiently small size, entropy can be computed from Boltzmann’s formula, the law of its temporal increase may have
to be replaced by an as yet unknown version of the law, perhaps
a statistical one. This is a fascinat ing and largely unexplored area; see refs. 3 and 4 for additional remarks.
“Extensivity,” or additivity, can break down on large scales as well, in the presence of gravitational interactions, as
mentioned by Lands berg. This is correct, well known, and worthy of emphasis. However, it lies outside the realm of
the labora tory physics under discussion.
Nussbaum’s remark about the Gibbs–Boltzmann approach being “the wrong direction” could be misin terpreted to
imply that we have an anti-Gibbs–Boltzmann bias. This would be unfortunate since we do value statistical mechanics
as does any physicist. We said, and we main tain, that the second law, as under stood for equilibrium states of
macroscopic systems, does not require statistical mechanics, or any other particular mechanics, for its existence. It does
require certain properties of macroscopic systems, and statistical mechanics is one model that, hopefully, can give those
properties, such as irreversibility. One should not confuse the exis tence, importance, and usefulness of the Boltzmann-
Gibbs-Maxwell theo ry with its necessity on the macro scopic level as far as the second law is concerned. Another way
to make the point is this: If the statistical mechanics of atoms is essential for the second law, then that law must imply
something about atoms and their dynamics. Does the second law prove the existence of atoms in the way that light
scattering, for exam ple, tells us what Avogadro’s number has to be? Does the law distinguish between classical and
quantum mechanics? The answer to these and similar questions is “No,” and if there were a direct connection, the late
19th-century wars about the existence of atoms would have been won much sooner. Alas, there is no such direct
connection that we are aware of, despite the many examples in which atomic constants make an appearance at the
macroscopic level— Planck’s radiation formula, the Sack ur–Tetrode equation, stability of matter with Coulomb forces,
and so on. The second law, however, is not such an example.
As our title read, this is a fresh look at the second law of thermody namics. It is no more obligatory than any other
approach, but happily some readers and colleagues have found it useful.
continued on page 106
Optimal Vision:
Blurring and Aliasing I n his article “Retinal Imaging and
Vision at the Frontiers of Adaptive Optics” (PHYSICS TODAY, January, page 31), Donald T. Miller shows that it is
possible to improve the res olution, contrast, and clarity of reti nal images by correcting for defects in the eye’s optics.
We agree that “the best retinal image quality is obtained with the largest physiological pupil diameter (8 mm) and with
full correc tion of all ocular aberrations.” Howev er, we disagree with Miller’s sugges tion that the quality of vision may
be improved similarly (to achieve “super normal vision”) if the eye’s optics could be “corrected” with “adaptive optics”
to produce the performance of an aberration-free 8-mm lens.
The angular spacing between reti nal photoreceptors, as Miller states, “represents a neural limitation to visual
resolution.” In terms of com munication theory, this spacing determines the sampling passband of the eye that,
analogous to the bandwidth of a communication chan nel, sets an upper bound on the high est spatial frequencies that
the eye can convey to the higher levels of the brain. The preferred modulation transfer function (MTF)—or spatial
frequency response—of the eye’s optics relative to this sampling pass band is inescapably a compromise between
blurring and aliasing. Because the MTF decreases smoothly with increasing frequency, aliasing can be substantially
decreased only at
the cost of blurring and vice versa. If blurring and aliasing are prop erly accounted for in terms of their effect on the
information rate that the eye conveys to the higher levels of the brain, then it is the MTF of the 3-mm lens rather than
that of the 8-mm lens that, in normal day light, maximizes this rate for the 50 cycles/degree sampling passband of the
eye. Hence, communication the ory and evolution converge, under appropriate conditions, toward the same optical
1
design. And why not? It seems unlikely that evolution would have missed the opportunity to improve our vision if it
could have done so merely by permitting the pupil to be wider than 3 mm during normal daylight conditions.
Reference
1. F. O. Huck, C. L. Fales, D. J. Jobson, Z. Rahman, Opt. Eng. 34, 795 (1995). F. O. Huck, C. L. Fales, Z. Rahman, Philos. Trans.
Roy. Soc. London, A354, 2193-2248 (1996). F. O. Huck, C. L. Fales, Z. Rahman, Visual Communica tion: An Information Theory
Approach, Kluwer Academic, Norwell, Mass., 1997. FRIEDRICH O. HUCK
(f.o.huck@larc.nasa.gov)
CARL L. FALES
NASA Langley Research Center Hampton, Virginia
Moore’s Law and the Future of Computing J oel Birnbaum and R. Stanley
Williams of Hewlett-Packard com ment on Moore’s law (P HYSICS TODAY, January, page 38) and dis cuss the
projections of its theoreti cally anticipated validity until 2012, or even 2020. However, it is more a matter of
practicable engineering and technology than it is of theoreti cal limits of the physical theory.
Interestingly, the researchers of Intel Corp see the whole develop ment rather more pessimistically. For instance,
David Papworth, Intel Fellow, suggests that Moore’s law won’t survive beyond 2004–5. He noted at a VLSi Circuits
Symposium in Hawaii in 1998 that by using two or three times as many transistors to increase performance 1.8 times,
progress continued apace but at cost: Power consumption has doubled or tripled in each generation. Papworth
concluded that, after 2004, density increases will slow down. 1