Skip to the Main Content

Note:These pages make extensive use of the latest XHTML and CSS Standards. They ought to look great in any standards-compliant modern browser. Unfortunately, they will probably look horrible in older browsers, like Netscape 4.x and IE 4.x. Moreover, many posts use MathML, which is, currently only supported in Mozilla. My best suggestion (and you will thank me when surfing an ever-increasing number of sites on the web which have been crafted to use the new standards) is to upgrade to the latest version of your browser. If that's not possible, consider moving to the Standards-compliant and open-source Mozilla browser.

June 14, 2024

100 Papers on Magnitude

Posted by Tom Leinster

A milestone! By my count, there are now 100 papers on magnitude, including several theses, by a total of 73 authors. You can find them all at the magnitude bibliography.

Here I’ll quickly do two things: tell you about some of the hotspots of current activity, then — more importantly — describe several regions of magnitude-world that haven’t received the attention they could have, and where there might even be some low-hanging fruit.

Where to start reading about magnitude

Incidentally, if you’re curious about magnitude but don’t know what to read first, one possible starting point is these colloquium slides. The basic shape of the theory — glossing over technicalities — is this:

  • Magnitude is a numerical invariant of enriched categories. It is arguably the canonical measure of the size of an enriched category, whatever “size” means.

  • Magnitude homology is an algebraic invariant of enriched categories. Indeed, it is arguably the canonical homology theory of enriched categories. It is also intended to be a categorification of magnitude.

  • Enriched categories include, among other things, ordinary categories and metric spaces. The magnitude of a category is often called its Euler characteristic. For metric spaces, magnitude and magnitude homology are something really new, so that’s where a lot of the interest lies.

For more detail, here’s a 2016 survey of magnitude, excluding magnitude homology:

Tom Leinster and Mark Meckes. The magnitude of a metric space: from category theory to geometric measure theory. arXiv:1606.00095; in Nicola Gigli (ed.), Measure Theory in Non-Smooth Spaces, de Gruyter Open, 2017.

Magnitude homology was first introduced for graphs:

Richard Hepworth and Simon Willerton. Categorifying the magnitude of a graph. arXiv:1505.04125; Homology, Homotopy and Applications 19(2) (2017), 31–60.

Graphs can be seen as special metric spaces, where the distance between two vertices it the number of edges in a shortest path between them. In turn, metric spaces are special enriched categories. The extension of magnitude homology from graphs to enriched categories was done here:

Tom Leinster and Michael Shulman. Magnitude homology of enriched categories and metric spaces. arXiv:1711.00802; Algebraic & Geometric Topology 21 (2021), 2175–2221.

Several people have pointed out that it would be really useful to have a modern survey of magnitude and magnitude homology. A bunch of us are working on it! But for now, these are my best suggestions.

Hotspots

Right now, it’s undoubtedly magnitude homology that’s receiving the most attention. The magnitude homology of graphs is especially active, although there’s been a lot on general metric spaces too.

Here are a few of the topics in this sphere that people have been investigating:

  • The relationship between magnitude homology and persistent homology of a metric space — which detect very different information about a space (Nina Otter; Simon Cho)

  • Applications of magnitude homology to the analysis of networks (Chad Giusti and Giuliamaria Menara)

  • A theory of magnitude cohomology (Richard Hepworth)

  • A comprehensive spectral sequence approach that encompasses both magnitude homology and path homology (Yasuhiko Asao; Richard Hepworth and Emily Roff; Kiyonori Gomi; Sergei Ivanov, …)

  • A concept of magnitude homotopy (Yu Tajima and Masahiko Yoshinaga)

  • An analysis of the inherent complexity of magnitude homology of graphs (Radmila Sazdanovic and Victor Summers; Luigi Caputi and Carlo Collari)

  • New results on magnitude homology equivalence of subsets of n\mathbb{R}^n, involving convex geometry (Adrián Doña Mateo and myself).

In need of attention

Although some topics in magnitude have received intense attention, there are others that remain wide open. Doubtless they won’t all be equally fruitful, but I think there are real opportunities to do new and exciting things. Some might not even be that hard.

Here’s my list, in no particular order.

  • The magnitude of graphs   The magnitude of a graph is an invariant taking values in the ring of rational functions over \mathbb{Q} in one variable. Or, from another viewpoint, it takes values in power series over \mathbb{Z}. From some angles it looks a bit like the Tutte polynomial, but neither is a specialization of the other.

    The magnitude homology of graphs is being thoroughly investigated, but magnitude itself not so much. As far as I know, it has received zero attention from graph theorists, or combinatorialists more generally, presumably because it hasn’t been used to solve an open problem they care about.

    I developed the theory myself as far as I could, but I’m not remotely skilled in graph theory. I’d love to see what the magnitude of graphs would look like in the hands of an expert graph theorist. What would they want to prove? What questions would they ask? What use could they put it to?

  • Effective sample size   In statistics, there’s a notion of effective sample size, which I learned of long ago from Paul Blackwell and blogged about here. And in a certain context, the effective sample size is the magnitude of the correlation matrix! That is, it’s the sum of all the entries of the inverse matrix.

    This is very likely not a coincidence, as the world of magnitude already involves concepts like “effective number of points” and “effective number of species”. But the connection between magnitude and effective sample size is almost completely uninvestigated.

  • The magnitude of higher categories   The way magnitude works is that given a notion of the size of each object of some monoidal category VV, one automatically obtains a notion of the size of any finite category enriched in VV. For instance, starting with the notion of the cardinality of a finite set, one automatically obtains the notion of the magnitude of a finite category.

    And we can keep going! From the notion of the magnitude of a finite category, we automatically get a notion of the magnitude of a finite 2-category, since a 2-category is a category enriched in Cat\mathbf{Cat}. And so on, to get a notion of the magnitude of a finite nn-category.

    This appears to be a reasonable notion. For instance, there’s an nn-category S n\mathbf{S}^n that looks like the nn-sphere (e.g. S 1=()\mathbf{S}^1 = (\bullet \rightrightarrows \bullet)), and its magnitude is the Euler characteristic of the topological nn-sphere S nS^n (either 00 or 22, depending on the parity of nn). But apart from this single example and one paper by Kohei Tanaka on bicategories, there has barely been any research at all on the magnitude of nn-categories. And the question of the correct definition of the magnitude of a finite \infty-category is still wide open.

  • The magnitude of linear categories   By a “linear category”, I mean a category enriched in vector spaces. As in the previous bullet point, the notion of the dimension of a finite-dimensional vector space automatically gives rise to the notion of the magnitude of a finite linear category.

    Probably there are interesting theorems to be proved about the magnitude of linear categories. But as far as I know, there’s very little work on this so far: just the two papers listed in the magnitude bibliography. What we now know about magnitude homology may be especially useful here.

  • Spectral geometry and semiclassical analysis   Some of the most analytically sophisticated work on the magnitude of subsets of n\mathbb{R}^n has used techniques from these subjects. For example, this is how it was shown that for nice enough sets X,Y nX, Y \subseteq \mathbb{R}^n, an asymptotic inclusion-exclusion principle holds:

    mag(t(XY))+mag(t(XY))mag(tX)mag(tY)0 mag(t(X \cup Y)) + mag(t(X \cap Y)) - mag(t X) - mag(t Y) \to 0

    as tt \to \infty. Here tXt X is the space XX scaled by a factor of tt.

    Heiko Gimperlein, who did this work with Magnus Goffeng and Nikoletta Louca, has expressed the view that a lot more about magnitude could be proved with spectral geometry techniques, if only enough people with that expertise knew about the open problems and put the time in. He and his collaborators have already done a lot, but they’re only three people!

    As an example of what could be achieved by mathematicians with these skills, see the next bullet point…

  • Maximum entropy   Off to the edge of magnitude-world are the almost-synonymous notions of entropy and diversity. In particular, there is a notion of the entropy of a probability distribution on a compact metric space.

    Given any compact metric space XX, one can ask what the maximum entropy of XX is — that is, the maximum among the entropies of all possible probability distributions on XX. (Emily Roff, Mark Meckes and I established that such a thing is well-defined.) Like magnitude, this is a real number associated with any compact metric space. One can also ask which distribution on XX achieves that maximum.

    It turns out that maximum entropy is closely related to magnitude. (For example, the maximum entropy of XX is equal to the magnitude of the support of the maximizing distribution.) That’s how it fits into the world of magnitude.

    So: maximum entropy is a numerical invariant of metric spaces. But almost no examples are known! For instance, apart from the trivial one-dimensional case, no one knows the maximum entropy of a Euclidean ball — not in any dimension, not of any radius. No one knows what the maximizing distribution on it is either. No one even knows the support of the maximizing distribution. And if that’s the sorry state of knowledge about balls, you can imagine how ignorant we are about more complex spaces.

    I understand from Heiko that it should be possible to apply the analytic techniques mentioned in the previous bullet point to solve this kind of problem. But so far, no one’s tried.

  • The categorification problem   The final item on my list is different. It’s not so much an overlooked or understudied area as a really important problem that hasn’t been solved.

    Magnitude homology is intended to be a categorification of magnitude. There is a theorem making this precise for finite metric spaces, as follows.

    Magnitude homology of metric spaces XX is a real-graded homology theory: there is one group H n,(X)H_{n, \ell}(X) for each integer n0n \geq 0 and real 0\ell \geq 0. This means that for each real 0\ell \geq 0, we have an Euler characteristic

    χ (X)= n(1) nrank(H n,(X)). \chi_\ell(X) = \sum_n (-1)^n rank(H_{n, \ell}(X)).

    Now assemble these Euler characteristics into a single generating function in a formal variable qq:

    +χ (X)q \sum_{\ell \in \mathbb{R}^+} \chi_\ell(X) q^\ell

    The categorification result is this: when XX is finite, this generating function evaluated at q=e tq = e^{-t} gives exactly the magnitude of tXt X, for all t>0t \gt 0.

    But that only works for finite spaces. For infinite spaces, nothing like this can possibly true, at least with the current definitions. For example, the magnitude homology of a convex subset of n\mathbb{R}^n is trivial in homological degrees n>0n \gt 0, which means we can’t possibly recover anything from it. On the other hand, convex sets have very different magnitudes: among other things, the function tmag(tX)t \mapsto mag(t X) determines the volume and dimension of XX. So it’s certainly impossible to recover the magnitude of a convex set from its magnitude homology.

    The moral: something is wrong in the foundations of magnitude homology!

    The challenge is to fix it, in such a way that the categorification theorem extends from finite to infinite metric spaces. No one knows how.

In this post I’ve just listed my subjective impressions of which areas are particularly active in the world of magnitude, which are most in need of attention, and where I think there might be exciting new results to be found. I’m sure others would have different lists, and I’d probably write different ones myself if I did this tomorrow rather than today. I hope no one will be offended by me mentioning A rather than B, and I also hope that others with other ideas will put them in the comments.

Happy magnituding! Here’s to the next 100.

Posted at June 14, 2024 11:04 PM UTC

TrackBack URL for this Entry:   https://golem.ph.utexas.edu/cgi-bin/MT-3.0/dxy-tb.fcgi/3538

9 Comments & 0 Trackbacks

Re: 100 Papers on Magnitude

Fascinating to have seen the idea grow so far in so many directions.

Given

A theory of magnitude cohomology (Richard Hepworth)

and

A concept of magnitude homotopy (Yu Tajima and Masahiko Yoshinaga),

and given

The relationship between magnitude homology and persistent homology of a metric space,

then perhaps there’s a hope for

a magnitude cohomotopy and its relationship with persistent cohomotopy.

Posted by: David Corfield on June 18, 2024 3:56 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

Thanks, David!

That’s a nice idea. I think it’s well beyond what anyone is doing now. Homotopy in a magnitude context is still in a very embryonic state, and I’m pretty sure no one has got to cohomotopy. But why not?

Incidentally, the joint paper with Adrián Doña Mateo that I just blogged about today also dips its toes into homotopical waters. From one point of view, the metric analogue of ordinary homotopy between two continuous maps is that our maps f,g:XYf,g: X \to Y are equal on the inner boundary of XX. At least, that’s the case when XX and YY are something like subsets of Euclidean space.

This analogy is developed in 7.4-7.8 of our paper, with a crucial limitation of the analogy mentioned in Remark 7.8.

Posted by: Tom Leinster on June 18, 2024 4:10 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

I think I’m right in saying that magnitude (ESS really) approximately counts the number of distinct “blobs” for reasonably smooth Gaussian random fields. I said something about it on Mastodon.

For “rough” looking fields it’s a bit messier but I think there’s still a way to make sense of it.

Posted by: Dan Piponi on October 28, 2024 12:12 AM | Permalink | Reply to this

Re: 100 Papers on Magnitude

That thread looks super interesting! Thanks for linking to it. For others reading, ESS = effective sample size, which I fumbled my way through explaining here.

The thought has been floating around for a while that there’s some connection between magnitude and clusters, but it looks like you’ve thought about it in quite some detail. I hope to get my head around what you’ve done.

Posted by: Tom Leinster on October 28, 2024 9:20 AM | Permalink | Reply to this

Re: 100 Papers on Magnitude

A quick rewrite as I don’t like making use of mathstodon’s support for TeX:

Suppose F(x)F(x) is a random field, which you can think of as being an infinite number of random variables, indexed by points xx in some domain DD.

(For generality, DD could be of any dimension nn so x,y,x, y,\ldots will typically be nn-dimensional points.)

Let’s say the values of the field have a standard normal distribution F(x)~N(0,1)F(x)~N(0,1) but that they’re not independent so we have covariance E[F(x)F(y)]=k(xy)E[F(x)F(y)]=k(x-y) for some kk. (We must have k(0)=1k(0)=1.)

If we want to estimate E[F(x)]E[F(x)], then under reasonable conditions with suitably large DD the best estimate will be the average 1V DF(x)dx\frac{1}{V}\int_D F(x)dx where VV is the “volume” of DD.

We compute the variance of this estimate as:

E[(1V DF(x)dx) 2] =1V 2E[ D DF(x)F(y)dxdy] =1V 2 D Dk(xy)dxdy 1V Dk(x)dx \begin{aligned} E[(\frac{1}{V}\int_D F(x)dx)^2]&=\frac{1}{V^2}E[\int_D\int_D F(x)F(y)dxdy]\\ &=\frac{1}{V^2}\int_D\int_D k(x-y)dxdy\\ &\approx\frac{1}{V}\int_D k(x)dx\\ \end{aligned}

(Whenever those steps make sense. The approximate equality will hold for k(x)k(x) that decay rapidly enough to zero and large DD where “edge effects” of the boundary of DD become small compared to the volume inside.)

So we have an ESS of V/ Dk(x)dxV/\int_D k(x)dx.

I think this makes intuitive sense. The ESS is going to be proportional to the volume of the region you consider. k(x)k(x) measures the “region of influence” where points are correlated and the integral is sort of a measure of its size. You want to get out of the region of influence to find another field value that contributes a new independent data point to your estimate. And similarly, for nice smooth k(x)k(x) I’d expect the region of influence to correspond roughly to blobs because you recognise a blob as a region where the value is correlated.

BTW my original motivation for this was thinking about the “number” of “independent” speckles in laser speckle. This is all probably a standard thing in a textbook.

Posted by: Dan Piponi on November 1, 2024 6:27 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

Thanks! I’m trying to get my head round this, in snatched moments when I’m not cranking out set theory course materials.

The estimate for the variance of your estimate (!) is certainly reminiscent of the kind of magnitude approximation that Simon Willerton was doing in the early days.

I like the phrase “region of influence”.

Posted by: Tom Leinster on November 6, 2024 9:35 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

To relate this to more recent things, the variance of that estimator is exactly 1/D 2 k(λ D)1/D_2^k(\lambda_D) in the notation of Tom’s paper with Emily, where D 2 kD_2^k denotes the “diversity of order 2” with respect to the similarity kernel kk, and λ D\lambda_D is normalized Lebesgue measure on DD.

If μ\mu is any probability measure on DD, then by the same argument, the estimator F(x)dμ(x)\int F(x) d\mu(x) has variance 1/D 2 k(μ)1/D_2^k(\mu), which is minimized when μ\mu is the maximizing distribution on DD.

That the maximizing distribution is approximately λ D\lambda_D, under appropriate conditions, is closely related to the proof in Tom and Emily’s paper that λ D\lambda_D is the “uniform distribution” on D nD \subseteq \mathbb{R}^n, in the sense defined in that paper.

Posted by: Mark Meckes on November 10, 2024 4:59 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

To make some connection with magnitude more explicit:

My remark above shows that the minimum variance of any estimator for 𝔼F(x)\mathbb{E}F(x) of a certain form is the reciprocal of the maximum diversity of DD. Thus, by another result in Tom and Emily’s paper, it is the reciprocal of a generalized notion of magnitude (defined using the kernel kk in place of the exponential kernel we usually use) of some subset of DD.

For a discrete parameter set DD, this means that minimum variance is the reciprocal of the magnitude of some submatrix of the covariance matrix for the family of random variables F(x)F(x). Which is very close to what Tom wrote about in his blog post on effective sample size. I haven’t sorted out yet whether I’m essentially just reproducing what Tom wrote about there, or generalizing it, or what.

Posted by: Mark Meckes on November 10, 2024 11:23 PM | Permalink | Reply to this

Re: 100 Papers on Magnitude

Going back and rereading Tom’s old post, what Tom wrote about there amounted to working with a finite parameter set DD, but allowing signed measures μ\mu with total mass 11.

There’s certainly a continuous version here as well: the minimum sample variance of an estimator of the form F(x)dμ(x)\int F(x) d\mu(x) for μ\mu a signed measure with total mass 11 is the reciprocal of the “kk-magnitude” of DD. This presumably requires some hypotheses on μ\mu and FF, and all the necessary tools should be in Tom and Emily’s paper, and maybe my “Positive definite metric spaces” paper.

The next question is, is this interesting enough to anyone to be worth pounding out the details?

Posted by: Mark Meckes on November 11, 2024 8:33 PM | Permalink | Reply to this

Post a New Comment