Potential Functions and the Magnitude of Functors 1
Posted by Tom Leinster
In the beginning, there were hardly any spaces whose magnitude we knew. Line segments were about the best we could do. Then Mark Meckes introduced the technique of potential functions for calculating magnitude, which was shown to be very powerful. For instance, Juan Antonio Barceló and Tony Carbery used it to compute the magnitude of odd-dimensional Euclidean balls, which turn out to be rational functions of the radius. Using potential functions allows you to tap into the vast repository of knowledge of PDEs.
In this post and the next, I’ll explain this technique from a categorical viewpoint, saying almost nothing about the analytic details. This is category theory as an organizational tool, used to help us understand how the various ideas fit together. Specifically, I’ll explain potential functions in terms of the magnitude of functors, which I wrote about here a few weeks ago.
Before I can describe this categorical viewpoint on potential functions, I have to explain what potential functions are in the magnitude context, and why they’re very useful. That’s what I’ll do today.
This part of the story is about metric spaces. For now I’ll assume they satisfy all the classical axioms, including symmetry of the metric, meaning that for all points and . When metric spaces are viewed as enriched categories, symmetry isn’t automatic — but we’ll come to that next time.
A weighting on a finite metric space is a function such that for all ,
Everyone who sees this formula for the first time asks where the exponential comes from. Ultimately it’s because of the enriched category viewpoint (which again we’ll come to next time), but the short story is that exponential is essentially the only reasonable function that converts addition into multiplication.
For simplicity, I’ll assume here that every finite metric space has a unique weighting, which I’ll call . Since the definition of weighting involves the same number of equations as unknowns, this is generically true (and it’s always true for subspaces of ), even though there are exceptions.
The magnitude of is
That’s for finite metric spaces. To extend the definition to compact metric spaces, there are various ideas you might try. You could define the magnitude of a compact space as the supremum of all the magnitudes of its finite subspaces. Or, you could take an ever-denser sequence of finite subsets of your compact space, then define its magnitude to be the limit of the magnitudes of the approximating subsets. Or, you could try replacing the sums in the formulas above by integrals, somehow.
Mark Meckes showed that all these approaches are equivalent. They all give the same definition of the magnitude of a compact space. (At least, this is true subject to a condition called “positive definiteness” which I won’t discuss and which always holds for subspaces of .)
How do we actually calculate magnitude, say for compact subspaces of ?
In principle, you already know how do it. You run through all the finite subsets of , calculate the magnitude of each using the definition above, and take the sup. The trouble is, this procedure is incredibly hard to work with. It’s completely impractical.
A slightly more practical approach is to look for a “weight measure”, that is, a signed measure on such that for all ,
As Mark showed, if is a weight measure then the magnitude of is given by . This is the analogue of the formula for finite spaces.
Example Take an interval in . It’s an elementary exercise to check that
is a weight measure on , where and denote the Dirac deltas at and , and is Lebesgue measure on . It follows that is the total mass of this measure, which is
In other words, it’s plus half the length of the interval.
The trouble with weight measures is that very few spaces have them. Even Euclidean balls don’t, beyond dimension one. It turns out that we need something more general than a measure, something more like a distribution.
In another paper, Mark worked out exactly what kind of measure-like or distribution-like thing a weighting should be. The answer is very nice, but I won’t explain it here, because I want to highlight the formal aspects above all.
So I’ll simply write the weighting on a metric space as , without worrying too much about what kind of thing actually is. All that matters is that it behaves something like a measure: we can pair it with “nice enough” functions to get a real number, which I’ll write as
or simply
And I’ll assume that all the spaces that we discuss do have a unique weighting .
That’s the background. But back to the question: how do we actually calculate magnitude?
First main idea Don’t look at metric spaces in isolation. Instead, consider spaces embedded in some big space that we understand well.
Think of the big space as fixed. The typical choice is . One sense in which we “understand” well is that we have a weight measure on it: it’s just Lebesgue measure divided by a constant factor . This follows easily from the fact that is homogeneous. (It doesn’t matter for anything I’m going to say, but the constant is , for which there’s a standard formula involving s and factorials.)
The potential function of a subspace is the function defined by
By definition of weighting, has constant value on . But outside , it could be anything, and it turns out that its behaviour outside is very informative.
Although is a measure-like thing on the subspace of , we typically extend it by to all of , and then the definition of the potential function becomes
Examples In all these examples, I’ll take the big space to be .
Let . Then the weighting on is the Dirac delta at , and the potential function is given by
(), whose graph looks like this:
Let . As we’ve already seen,
and a little elementary work then reveals that the potential function is
So the potential function looks like this:
In the two examples we just did, the potential function of is just the negative exponential of the distance between and the argument . That’s not exactly coincidence: as we’ll see in the next part, the function just described corresponds to a strict colimit, whereas the potential function corresponds to a lax colimit. So it’s not surprising that they coincide in simple cases.
But this only happens in the very simplest cases. It doesn’t happen for Euclidean balls above dimension , or even for two-point spaces. For example, taking the subset of , we have
which looks like this:
whereas looks like this:
Similar, but different!
And now we come to the:
Second main idea Work with potential functions instead of weightings.
To explain what this means, I need to tell you three good things about potential functions.
The potential function determines the magnitude:
At the formal level, proving this is a one-line calculation: substitute the definition of into the right-hand side and follow your nose.
For example, we saw that has weighting , where is Lebesgue measure and is a known constant. So
for compact . Here refers to ordinary Lebesgue integration.
(For , and in fact a bit more generally, this result appears as Theorem 4.16 here.)
You can recover the weighting from the potential function. So, you don’t lose any information by working with one rather than the other.
How do you recover it? Maybe it’s easiest to explain in the case when the spaces are finite. If we write for the matrix then the definition of the potential function can be expressed very succinctly:
Here and are viewed as column vectors with entries indexed by the points of . (For , those entries are for points not in ). Assuming is invertible, this means we recover from as . And something similar is true in the non-finite setting.
However, what really makes the technique of potential functions sing is that when , there’s a much more explicit way to recover the weighting from the potential function:
(up to a constant factor that I’ll ignore). Here is the identity and is the Laplace operator, . This is Proposition 5.9 of Mark’s paper.
How much more of this bullet point you’ll want to read depends on how interested you are in the analysis. The fundamental point is simply that is some kind of differential operator. But for those who want a bit more:
To make sense of everything, you need to interpret it all in a distributional sense. In particular, this allows one to make sense of the power , which is not an integer if is even.
Maybe you wondered why someone might have proved a result on the magnitude of odd-dimensional Euclidean balls only, as I mentioned at the start of the post. What could cause the odd- and even-dimensional cases to become separated? It’s because whether or not is an integer depends on whether is odd or even. When is odd, it’s an integer, which makes a differential rather than pseudodifferential operator. Heiko Gimperlein, Magnus Goffeng and Nikoletta Louca later worked out lots about the even-dimensional case, but I won’t talk about that here.
Finally, where does the operator come from? Sadly, I don’t have an intuitive explanation. Ultimately it comes down to the fact that the Fourier transform of is (up to a constant). But that itself is a calculation that’s really quite tricky (for me), and it’s hard to see anything beyond “it is what it is”.
The third good thing about potential functions is that they satisfy a differential equation, in the situation where our big space is . Specifically:
Indeed, the definition of weighting implies that on , and the “second good thing” together with the fact that is supported on give the second clause.
Not only do we have a differential equation for , we also have boundary conditions. There are boundary conditions at the boundary of , because of something I’ve been entirely vague about: the functions we’re dealing with are meant to be suitably smooth. There are also boundary conditions at , because our functions are also meant to decay suitably fast.
Maybe Mark will read this and correct me if I’m wrong, but I believe there are exactly the right number of boundary conditions to guarantee that there’s (typically? always?) a unique solution. In any case, the following example — also taken from Mark’s paper — illustrates the situation.
Example Let’s calculate the magnitude of a real interval using the potential function method.
Its potential function is a function such that on and on the rest of the real line. (That expression comes from taking in the case .)
The functions satisfying are those of the form for some constant , and in our case we’re choose the constant and the sign differently on the two connected components of . So there are lots of solutions. But is required to be continuous and to converge to at , and that pins it down uniquely: the one and only solution is
I showed you the graph of this potential function above, in the case where .
So the magnitude of is
where is the constant . This gives the answer:
The crucial point is that in this example, we didn’t have to come up with a weighting on . The procedure was quite mechanical. And that’s the attraction of the method of potential functions.
Next time, I’ll put all this into a categorical context using the notion of the magnitude of a functor, which I introduced here.