The Space of Physical Frameworks (Part 2)
Posted by John Baez
I’m trying to work out how classical statistical mechanics can reduce to thermodynamics in a certain limit. I sketched out the game plan in Part 1 but there are a lot of details to hammer out. While I’m doing this, let me stall for time by explaining more precisely what I mean by ‘thermodynamics’. Thermodynamics is a big subject, but I mean something more precise and limited in scope.
Thermostatic systems
A lot of what we call ‘thermodynamics’, or more precisely ‘classical thermodynamics’, has nothing to do with dynamics. It’s really about systems in equilibrium, not changing, so it actually deserves to be called ‘thermostatics’. Here’s one attempt to formalize a core idea:
Definition. A thermostatic system is a convex space together with a concave function . We call the space of states, and call the entropy of the state .
There’s a lot packed into this definition:
- The general concept of convex space: it’s roughly a set where you can take convex combinations of points , like where .
- How we make into a convex space: it’s pretty obvious, except that beats in convex combinations, like .
- What is a ‘concave’ function : it’s a function with
To see all the details spelled out with lots of examples, try this:
- John Baez, Owen Lynch and Joe Moeller, Compositional thermostatics, J. Math. Phys. 64 (2023) 023304. (Blog articles here.)
We actually defined a category of thermostatic systems and maps between them.
What you can do with a thermostatic system
For now I will only consider thermostatic systems where , made into a convex set in the usual way. In these examples a state is solely determined by its energy . I’m trying to keep things as simple as possible, and generalize later only if my overall plan actually works.
Here’s what people do in this very simple setting. Our thermostatic system is a concave function
describing the entropy of our system when it has energy . But often entropy is also a strictly increasing function of energy, with as . In this case, it’s impossible for a system to literally maximize entropy. What it does instead is maximize ‘entropy minus how much it spends on energy’ — just as you might try to maximize the pleasure you get from eating doughnuts minus your displeasure at spending money. Thus, if is the ‘cost’ of energy, our system tries to maximize
The cost is the reciprocal of a quantity called temperature:
So, should be called inverse temperature, and the rough intuition you should have is this. When it’s hot, energy is cheap and our system’s energy can afford to be high. When it’s cold, energy costs a lot and our system will not let its energy get too high.
If as a function of is differentiable and has a maximum, the maximum must occur at a point where
or
This gives the fundamental relation between energy, entropy and temperature:
However, the math will work better for us if we use the inverse temperature.
Suppose we have a system maximizing for some value of . The maximum value of is called free entropy and denoted . In short:
or if you prefer
This way of defining in terms of is called a Legendre–Fenchel transform, though conventions vary about the precise definition of this transform, and also its name. Since I’m lazy, I’ll just call it the Legendre transform. For more, read this:
The great thing about Legendre transforms is that if a function is convex and lower semicontinuous, when you take its Legendre transform twice you get that function back! This is part of the Fenchel–Moreau theorem. So under these conditions we automatically get another formula that looks very much like the one we’ve just seen:
When has a minimum as a function of and it’s differentiable there, this minimum must occur at a point where
or
Summary
I’m plotting a difficult course between sticking with historical conventions in thermodynamics and trying to make everything mathematically elegant. Everything above looks more elegant if we work with minus the free entropy, . Starting from a thermostatic system we then get a beautifully symmetrical pair of relations:
If the first infimum is achieved at some energy and is differentiable there, then
at this value of , and this formula lets us compute the inverse temperature as a function of . Similarly, if the second infimum is achieved at some and is differentiable there, then
at this value of , and this formula lets us compute as a function of .
When we describe a thermostatic system as a limit of classical statistical mechanical systems, these are the formulas we’d like to see emerge in the limit!
Appendix: the traditional formalism
If you’ve never heard of ‘free entropy’, you may be relieved to hear it’s a repackaging of the more familiar concept of ‘free energy’. The free energy , or more specifically the Helmholtz free energy, is related to the free entropy by
Unless you’re a real die-hard fan of thermodynamics, don’t read the following stuff, since it will only further complicate the picture I’ve tried to paint above, which is already blemished by the fact that physicists prefer to . I will not provide any profound new insights: I will merely relate what I’ve already explained to an equivalent but more traditional formalism.
I’ve been treating entropy as a function of energy: this is the so-called entropy scheme. But it’s traditional to treat energy as a function of entropy: this is called the energy scheme.
The entropy scheme generalizes better. In thermodynamics we often want to think about situations where entropy is a function of several variables: energy, volume, the amounts of various chemicals, and so on. Then we should work with a thermostatic system where is a convex subset of . Everything I did generalizes nicely to that situation, and now will be one of quantities that arise by taking a Legendre transform of .
But when entropy is a function of just one variable, energy, people often turn the tables and try to treat energy as a function of entropy, say . They then define the free energy as a function of temperature by
This is essentially a Legendre transform — but notice that inside the parentheses we have instead of . We can fix this by using a sup instead of an inf, and writing
It’s actually very common to define the Legendre transform using a sup instead of an inf, so that’s fine. The only wrinkle is that this Legendre transform gives us instead of .
When the supremum is achieved at a point where is differentiable we have
at that point. When is concave and lower semicontinuous, taking its Legendre transform twice gets us back where we started:
And when this supremum is achieved at a point where is differentiable, we have
To top it off, physicists tend to assume and take values where the suprema above are achieved, and not explicitly write what is a function of what. So they would summarize everything I just said with these equations:
If instead we take the approach I’ve described, where entropy is treated as a function of energy, it’s natural to focus on the negative free entropy and inverse temperature . If we write the equations governing these in the same slapdash style as those above, they look like this:
Less familiar, but more symmetrical! The two approaches are related by
Thermodynamics is a funny subject. The first time you go through it, you don’t understand it at all. The second time you go through it, you think you understand it, except for one or two points.The third time you go through it, you know you don’t understand it, but by that time you are so used to the subject, it doesn’t bother you anymore. — Arnold Sommerfeld
Re: The Space of Physical Frameworks (Part 2)
If you had the misfortune to read the above post before I wrote this comment here, please know that it had some serious mistakes which I’ve fixed now. I’ve also added a lot to the final section, including a funny quote at the end.