Planet Musings

December 10, 2025

Terence TaoThe Equational Theories Project: Advancing Collaborative Mathematical Research at Scale

Matthew Bolan, Joachim Breitner, Jose Brox, Nicholas Carlini, Mario Carneiro, Floris van Doorn, Martin Dvorak, Andrés Goens, Aaron Hill, Harald Husum, Hernán Ibarra Mejia, Zoltan Kocsis, Bruno Le Floch, Amir Livne Bar-on, Lorenzo Luccioli, Douglas McNeil, Alex Meiburg, Pietro Monticone, Pace P. Nielsen, Emmanuel Osalotioman Osazuwa, Giovanni Paolini, Marco Petracci, Bernhard Reinke, David Renshaw, Marcus Rossel, Cody Roux, Jérémy Scanvic, Shreyas Srinivas, Anand Rao Tadipatri, Vlad Tsyrklevich, Fernando Vaquerizo-Villar, Daniel Weber, Fan Zheng, and I have just uploaded to the arXiv our preprint The Equational Theories Project: Advancing Collaborative Mathematical Research at Scale. This is the final report for the Equational Theories Project, which was proposed in this blog post and also showcased in this subsequent blog post. The aim of this project was to see whether one could collaboratively achieve a large-scale systematic exploration of a mathematical space, which in this case was the implication graph between 4694 equational laws of magmas. A magma is a set {G} equipped with a binary operation {\diamond: G \times G \rightarrow G} (or, equivalently, a {G \times G} multiplication table). An equational law is an equation involving this operation and a number of indeterminate variables. Some examples of equational laws, together with the number that we assigned to that law, include

Up to relabeling and symmetry, there turn out to be 4694 equational laws that involve at most four invocations of the magma operation {\diamond}; one can explore them in our “Equation Explorer” tool.

The aim of the project was to work out which of these laws imply which others. For instance, all laws imply the trivial law {E1}, and conversely the singleton law {E2} implies all the others. On the other hand, the commutative law {E43} does not imply the associative law {E4512} (because there exist magmas that are commutative but not associative), nor is the converse true. All in all, there are {22,028,942} implications of this type to settle; most of these are relatively easy and could be resolved in a matter of minutes by an expert in college-level algebra, but prior to this project, it was impractical to actually do so in a manner that could be feasibly verified. Also, this problem is known to become undecidable for sufficiently long equational laws. Nevertheless, we were able to resolve all the implications informally after two months, and have them completely formalized in Lean after a further five months.

After a rather hectic setup process (documented in this personal log), progress came in various waves. Initially, huge swathes of implications could be resolved first by very simple-minded techniques, such as brute-force searching all small finite magmas to refute implications; then, automated theorem provers such as Vampire or Mace4 / Prover9 were deployed to handle a large fraction of the remainder. A few equations had existing literature that allowed for many implications involving them to be determined. This left a core of just under a thousand implications that did not fall to any of the “cheap” methods, and which occupied the bulk of the efforts of the project. As it turns out, all of the remaining implications were negative; the difficulty was to construct explicit magmas that obeyed one law but not another. To do this, we discovered a number of general constructions of magmas that were effective at this task. For instance:

  • Linear models, in which the carrier {G} was a (commutative or noncommutative) ring and the magma operation took the form {x \diamond y = ax + by} for some coefficients {a,b}, turned out to resolve many cases.
  • We discovered a new invariant of an equational law, which we call the “twisting semigroup” of that law, which also allowed us to construct further examples of magmas that obeyed one law {E} but not another {E'}, by starting with a base magma {M} that obeyed both laws, taking a Cartesian power {M^n} of that magma, and then “twisting” the magma operation by certain permutations of {\{1,\dots,n\}} designed to preserve {E} but not {E'}.
  • We developed a theory of “abelian magma extensions”, similar to the notion of an abelian extension of a group, which allowed us to flexibly build new magmas out of old ones in a manner controlled by a certain “magma cohomology group {H^1_E(G,M)}” which were tractable to compute, and again gave ways to construct magmas that obeyed one law {E} but not another {E'}.
  • Greedy methods, in which one fills out an infinite multiplication table in a greedy manner (somewhat akin to naively solving a Sudoku puzzle), subject to some rules designed to avoid collisions and maintain a law {E}, as well as some seed entries designed to enforce a counterexample to a separate law {E'}. Despite the apparent complexity of this method, it can be automated in a manner that allowed for many outstanding implications to be refuted.
  • Smarter ways to utilize automated theorem provers, such as strategically adding in additional axioms to the magma to help narrow the search space, were developed over the course of the project.

Even after applying all these general techniques, though, there were about a dozen particularly difficult implications that resisted even these more powerful methods. Several ad hoc constructions were needed in order to understand the behavior of magmas obeying such equations as E854, E906, E1323, E1516, and E1729, with the latter taking months of effort to finally solve and then formalize.

A variety of GUI interfaces were also developed to facilitate the collaboration (most notably the Equation Explorer tool mentioned above), and several side projects were also created within the project, such as the exploration of the implication graph when the magma was also restricted to be finite. In this case, we resolved all of the {22,028,942} implications except for one (and its dual):

Problem 1 Does the law {x = y \diamond (x \diamond ((y \diamond x) \diamond y))} (E677) imply the law {x = ((x \diamond x) \diamond x) \diamond x} (E255) for finite magmas?

See this blueprint page for some partial results on this problem, which we were unable to resolve even after months of effort.

Interestingly, modern AI tools did not play a major role in this project (but it was largely completed in 2024, before the most recent advanced models became available); while they could resolve many implications, the older “good old-fashioned AI” of automated theorem provers were far cheaper to run and already handled the overwhelming majority of the implications that the advanced AI tools could. But I could imagine that such tools would play a more prominent role in future similar projects.

Terence TaoThe story of Erdős problem #1026

Problem 1026 on the Erdős problem web site recently got solved through an interesting combination of existing literature, online collaboration, and AI tools. The purpose of this blog post is to try to tell the story of this collaboration, and also to supply a complete proof.

The original problem of Erdős, posed in 1975, is rather ambiguous. Erdős starts by recalling his famous theorem with Szekeres that says that given a sequence of {k^2+1} distinct real numbers, one can find a subsequence of length {k+1} which is either increasing or decreasing; and that one cannot improve the {k^2+1} to {k^2}, by considering for instance a sequence of {k} blocks of length {k}, with the numbers in each block decreasing, but the blocks themselves increasing. He also noted a result of Hanani that every sequence of length {k(k+3)/2} can be decomposed into the union of {k} monotone sequences. He then wrote “As far as I know the following question is not yet settled. Let {x_1,\dots,x_n} be a sequence of distinct numbers, determine

\displaystyle  S(x_1,\dots,x_n) = \max \sum_r x_{i_r}

where the maximum is to be taken over all monotonic sequences {x_{i_1},\dots,x_{i_m}}“.

This problem was added to the Erdős problem site on September 12, 2025, with a note that the problem was rather ambiguous. For any fixed {n}, this is an explicit piecewise linear function of the variables {x_1,\dots,x_n} that could be computed by a simple brute force algorithm, but Erdős was presumably seeking optimal bounds for this quantity under some natural constraint on the {x_i}. The day the problem was posted, Desmond Weisenberg proposed studying the quantity {c(n)}, defined as the largest constant such that

\displaystyle  S(x_1,\dots,x_n) \geq c(n) \sum_{i=1}^n x_i

for all choices of (distinct) real numbers {x_1,\dots,x_n}. Desmond noted that for this formulation one could assume without loss of generality that the {x_i} were positive, since deleting negative or vanishing {x_i} does not decrease the left-hand side and does not increase the right-hand side. By a limiting argument one could also allow collisions between the {x_i}, so long as one interpreted monotonicity in the weak sense.

Though not stated on the web site, one can formulate this problem in game theoretic terms. Suppose that Alice has a stack of {N} coins for some large {N}. She divides the coins into {n} piles of consisting of {x_1,\dots,x_n} coins each, so that {\sum_{i=1}^n x_i = N}. She then passes the piles to Bob, who is allowed to select a monotone subsequence of the piles (in the weak sense) and keep all the coins in those piles. What is the largest fraction {c(n)} of the coins that Bob can guarantee to keep, regardless of how Alice divides up the coins? (One can work with either a discrete version of this problem where the {x_i} are integers, or a continuous one where the coins can be split fractionally, but in the limit {N \rightarrow \infty} the problems can easily be seen to be equivalent.)

For small {n}, one can work this out by hand. For {n=1}, clearly {c(1)=1}: Alice has to put all the coins into one pile, which Bob simply takes. Similarly {c(2)=1}: regardless of how Alice divides the coins into two piles, the piles will either be increasing or decreasing, so in either case Bob can take both. The first interesting case is {n=3}. Bob can again always take the two largest piles, guaranteeing himself {2/3} of the coins. On the other hand, if Alice almost divides the coins evenly, for instance into piles {((1/3 + \varepsilon)N, (1/3-2\varepsilon) N, (1/3+\varepsilon)N)} for some small {\varepsilon>0}, then Bob cannot take all three piles as they are non-monotone, and so can only take two of them, allowing Alice to limit the payout fraction to be arbitrarily close to {2/3}. So we conclude that {c(3)=2/3}.

An hour after Desmond’s comment, Stijn Cambie noted (though not in the language I used above) that a similar construction to the one above, in which Alice divides the coins into {k^2} pairs that are almost even, in such a way that the longest monotone sequence is of length {k}, gives the upper bound {c(k^2) \leq 1/k}. It is also easy to see that {c(n)} is a non-increasing function of {n}, so this gives a general bound {c(n) \leq (1+o(1))/\sqrt{n}}. Less than an hour after that, Wouter van Doorn noted that the Hanani result mentioned above gives the lower bound {c(n) \geq (\frac{1}{\sqrt{2}}-o(1))/\sqrt{n}}, and posed the problem of determining the asymptotic limit of {\sqrt{n} c(n)} as {n \rightarrow \infty}, given that this was now known to range between {1/\sqrt{2}-o(1)} and {1+o(1)}. This version was accepted by Thomas Bloom, the moderator of the Erdős problem site, as a valid interpretation of the original problem.

The next day, Stijn computed the first few values of {c(n)} exactly:

\displaystyle  1, 1, 2/3, 1/2, 1/2, 3/7, 2/5, 3/8, 1/3.

While the general pattern was not yet clear, this was enough data for Stijn to conjecture that {c(k^2)=1/k}, which would also imply that {\sqrt{n} c(n) \rightarrow 1} as {n \rightarrow \infty}. (EDIT: as later located by an AI deep research tool, this conjecture was also made in Section 12 of this 1980 article of Steele.) Stijn also described the extremizing sequences for this range of {n}, but did not continue the calculation further (a naive computation would take runtime exponential in {n}, due to the large number of possible subsequences to consider).

The problem then lay dormant for almost two months, until December 7, 2025, in which Boris Alexeev, as part of a systematic sweep of the Erdős problems using the AI tool Aristotle, was able to get this tool to autonomously solve this conjecture {c(k^2)=1/k} in the proof assistant language Lean. The proof converted the problem to a rectangle-packing problem.

This was one further addition to a recent sequence of examples where an Erdős problem had been automatically solved in one fashion or another by an AI tool. Like the previous cases, the proof turned out to not be particularly novel. Within an hour, Koishi Chan gave an alternate proof deriving the required bound {c(k^2) \geq 1/k} from the original Erdős-Szekeres theorem by a standard “blow-up” argument which we can give here in the Alice-Bob formulation. Take a large {M}, and replace each pile of {x_i} coins with {(1+o(1)) M^2 x_i^2} new piles, each of size {(1+o(1)) x_i}, chosen so that the longest monotone subsequence in this collection is {(1+o(1)) M x_i}. Among all the new piles, the longest monotone subsequence has length {(1+o(1)) M S(x_1,\dots,x_n)}. Applying Erdős-Szekeres, one concludes the bound

\displaystyle  M S(x_1,\dots,x_n) \geq (1-o(1)) (\sum_{i=1}^{k^2} M^2 x_i^2)^{1/2}

and on canceling the {M}‘s, sending {M \rightarrow \infty}, and applying Cauchy-Schwarz, one obtains {c(k^2) \geq 1/k} (in fact the argument gives {c(n) \geq 1/\sqrt{n}} for all {n}).

Once this proof was found, it was natural to try to see if it had already appeared in the literature. AI deep research tools have successfully located such prior literature in the past, but in this case they did not succeed, and a more “old-fashioned” Google Scholar job turned up some relevant references: a 2016 paper by Tidor, Wang and Yang contained this precise result, citing an earlier paper of Wagner as inspiration for applying “blowup” to the Erdős-Szekeres theorem.

But the story does not end there! Upon reading the above story the next day, I realized that the problem of estimating {c(n)} was a suitable task for AlphaEvolve, which I have used recently as mentioned in this previous post. Specifically, one could task to obtain upper bounds on {c(n)} by directing it to produce real numbers (or integers) {x_1,\dots,x_n} summing up to a fixed sum (I chose {10^6}) with a small a value of {S(x_1,\dots,x_n)} as possible. After an hour of run time, AlphaEvolve produced the following upper bounds on {c(n)} for {1 \leq n \leq 16}, with some intriguingly structured potential extremizing solutions:

The numerical scores (divided by {10^6}) were pretty obviously trying to approximate simple rational numbers. There were a variety of ways (including modern AI) to extract the actual rational numbers they were close to, but I searched for a dedicated tool and found this useful little web page of John Cook that did the job:

\displaystyle  1, 1, 2/3, 1/2, 1/2, 3/7, 2/5, 3/8, 1/3, 1/4.

\displaystyle  1/3, 4/13, 3/10, 4/14, 3/11, 4/15, 1/4.

I could not immediately see the pattern here, but after some trial and error in which I tried to align numerators and denominators, I eventually organized this sequence into a more suggestive form:

\displaystyle  1,

\displaystyle  1/1, \mathbf{2/3}, 1/2,

\displaystyle  2/4, \mathbf{3/7}, 2/5, \mathbf{3/8}, 2/6,

\displaystyle  3/9, \mathbf{4/13}, 3/10, \mathbf{4/14}, 3/11, \mathbf{4/15}, 3/12.

which suggested a somewhat complicated but predictable conjecture for the values of the sequence {c(n)}. On posting this, Boris found a clean formulation of the conjecture, namely that

\displaystyle  c(k^2 + 2a + 1) = \frac{k}{k^2+a} \ \ \ \ \ (1)

whenever {k \geq 1} and {-k \leq a \leq k}. After a bit of effort, he also produced an explicit upper bound construction:

Proposition 1 If {k \geq 1} and {-k \leq a \leq k}, then {c(k^2+2a+1) \leq \frac{k}{k^2+a}}.

Proof: Consider a sequence {x_1,\dots,x_{k^2+2a+1}} of numbers clustered around the “red number” {|a|} and “blue number” {|a+1|}, consisting of {|a|} blocks of {k-|a|} “blue” numbers, followed by {|a+1|} blocks of {|a+1|} “red” numbers, and then {k-|a|} further blocks of {k} “blue” numbers, where all blocks are slightly increasing within each block, but the blue blocks are decreasing between each other, and the red blocks are decreasing between each other. The total number of elements is indeed

\displaystyle  |a| \times (k-|a|) + |a+1| \times |a+1| + (k-|a|) \times k

\displaystyle  = k^2 + 2a + 1

and the total sum is close to

\displaystyle |a| \times (k-|a|) \times |a+1| + |a+1| \times |a+1| \times |a|

+ (k-|a|) \times k \times |a+1| = (k^2 + a) |a+1|.

With this setup, one can check that any monotone sequence consists either of at most {|a+1|} red elements and at most {k-|a|} blue elements, or no red elements and at most {k} blue element, in either case giving a monotone sum that is bounded by either

\displaystyle  |a+1| \times |a| + (k-|a|) \times |a+1| = k |a+1|

or

\displaystyle  0 + k \times |a+1| = k |a+1|,

giving the claim. \Box

Here is a plot of 1/c(n) (produced by ChatGPT Pro), showing that it is basically a piecewise linear approximation to the square root function:

Shortly afterwards, Lawrence Wu clarified the connection between this problem and a square packing problem, which was also due to Erdős (Problem 106). Let {f(n)} be the least number such that, whenever one packs {n} squares of sidelength {d_1,\dots,d_n} into a square of sidelength {D}, with all sides parallel to the coordinate axes, one has

\displaystyle  \sum_{i=1}^n d_i \leq f(n) D.

Proposition 2 For any {n}, one has

\displaystyle  c(n) \geq \frac{1}{f(n)}.

Proof: Given {x_1,\dots,x_n} and {1 \leq i \leq n}, let {S_i} be the maximal sum over all increasing subsequences ending in {x_i}, and {T_i} be the maximal sum over all decreasing subsequences ending in {x_i}. For {i < j}, we have either {S_j \geq S_i + x_j} (if {x_j \geq x_i}) or {T_j \geq T_i + x_j} (if {x_j \leq x_i}). In particular, the squares {(S_i-x_i, T_i-x_i)} and {(S_j-x_j, T_j-x_j)} are disjoint. These squares pack into the square {[0, S(x_1,\dots,x_n)]^2}, so by definition of {f}, we have

\displaystyle  \sum_{i=1}^n x_i \leq f(n) S(x_1,\dots,x_n),

and the claim follows. \Box

This idea of using packing to prove Erdős-Szekeres type results goes back to a 1959 paper of Seidenberg, although it was a discrete rectangle-packing argument that was not phrased in such an elegantly geometric form. It is possible that Aristotle was “aware” of the Seidenberg argument via its training data, as it had incorporated a version of this argument in its proof.

Here is an illustration of the above argument using the AlphaEvolve-provided example [99998, 99997, 116305, 117032, 116304, 58370, 83179, 117030, 92705, 99080] for n=10 to convert it to a square packing (image produced by ChatGPT Pro):

At this point, Lawrence performed another AI deep research search, this time successfully locating a paper from just last year by Baek, Koizumi, and Ueoro, where they show that

Theorem 3 For any {k \geq 1}, one has

\displaystyle  f(k^2+1) \leq k

which, when combined with a previous argument of Praton, implies

Theorem 4 For any {k \geq 1} and {c \in {\bf Z}} with {k^2+2c+1 \geq 1}, one has

\displaystyle  f(k^2+2c+1) \leq k + \frac{c}{k}.

This proves the conjecture!

There just remained the issue of putting everything together. I did feed all of the above information into a large language model, which was able to produce a coherent proof of (1) assuming the results of Baek-Koizumi-Ueoro and Praton. Of course, LLM outputs are prone to hallucination, so it would be preferable to formalize that argument in Lean, but this looks quite doable with current tools, and I expect this to be accomplished shortly. But I was also able to reproduce the arguments of Baek-Koizumi-Ueoro and Praton, which I include below for completeness.

Proof: (Proof of Theorem 3, adapted from Baek-Koizumi-Ueoro) We can normalize {D=k}. It then suffices to show that if we pack the length {k} torus {({\bf Z}/k{\bf Z})^2} by {k^2+1} axis-parallel squares of sidelength {d_1,\dots,d_{k^2+1}}, then

\displaystyle  \sum_{i=1}^{k^2+1} d_i \leq k^2.

Pick {x_0, y_0 \in {\bf R}/k{\bf Z}}. Then we have a {k \times k} grid

\displaystyle  (x_0 + {\bf Z}) \times (y_0 + {\bf Z}) \pmod {k{\bf Z}^2}

inside the torus. The {i^{th}} square, when restricted to this grid, becomes a discrete rectangle {A_{i,x_0} \times B_{i,y_0}} for some finite sets {A_{i,x_0}, B_{i,y_0}} with

\displaystyle  |\# A_{i,x_0} -\# B_{i,y_0}| \leq 1. \ \ \ \ \ (2)

By the packing condition, we have

\displaystyle  \sum_{i=1}^{k^2+1} \# A_{i,x_0} \# B_{i,y_0} \leq k^2.

From (2) we have

\displaystyle  (\# A_{i,x_0} - 1) (\# B_{i,y_0} - 1) \geq 0

hence

\displaystyle  \# A_{i,x_0} \# B_{i,y_0} \geq \# A_{i,x_0} + \# B_{i,y_0} - 1.

Inserting this bound and rearranging, we conclude that

\displaystyle  \sum_{i=1}^{k^2+1} \# A_{i,x_0} + \sum_{i=1}^{k^2+1} \# B_{i,y_0} \leq 2k^2 + 1.

Taking the supremum over {x_0,y_0} we conclude that

\displaystyle  \sup_{x_0} \sum_{i=1}^{k^2+1} \# A_{i,x_0} + \sup_{y_0} \sum_{i=1}^{k^2+1} \# B_{i,y_0} \leq 2k^2 + 1

so by the pigeonhole principle one of the summands is at most {k^2}. Let’s say it is the former, thus

\displaystyle  \sup_{x_0} \sum_{i=1}^{k^2+1} \# A_{i,x_0} \leq k^2.

In particular, the average value of {\sum_{i=1}^{k^2+1} \# A_{i,x_0}} is at most {k^2}. But this can be computed to be {\sum_{i=1}^{k^2+1} d_i}, giving the claim. Similarly if it is the other sum. \Box

UPDATE: Actually, the above argument also proves Theorem 4 with only minor modifications. Nevertheless, we give the original derivation of Theorem 4 using the embedding argument of Praton below for sake of completeness.

Proof: (Proof of Theorem 4, adapted from Praton) We write {c = \epsilon |c|} with {\epsilon = \pm 1}. We can rescale so that the square one is packing into is {[0,k]^2}. Thus, we pack {k^2+2\varepsilon |c|+1} squares of sidelength {d_1,\dots,d_{k^2+2\varepsilon |c|+1}} into {[0,k]^2}, and our task is to show that

\displaystyle  \sum_{i=1}^{k^2+2\varepsilon|c|+1} d_i \leq k^2 + \varepsilon |c|.

We pick a large natural number {N} (in particular, larger than {k}), and consider the three nested squares

\displaystyle  [0,k]^2 \subset [0,N]^2 \subset [0,N + |c| \frac{N}{N-\varepsilon}]^2.

We can pack {[0,N]^2 \backslash [0,k]^2} by {N^2-k^2} unit squares. We can similarly pack

\displaystyle  [0,N + |c| \frac{N}{N-\varepsilon}]^2 \backslash [0,N]^2

\displaystyle  =[0, \frac{N}{N-\varepsilon} (N+|c|-\varepsilon)]^2 \backslash [0, \frac{N}{N-\varepsilon} (N-\varepsilon)]^2

into {(N+|c|-\varepsilon)^2 - (N-\varepsilon)^2} squares of sidelength {\frac{N}{N-\varepsilon}}. All in all, this produces

\displaystyle  k^2+2\varepsilon |c|+1 + N^2-k^2 + (N+|c|-\varepsilon)^2 - (N-\varepsilon)^2

\displaystyle   = (N+|c|)^2 + 1

squares, of total length

\displaystyle (\sum_{i=1}^{k^2+2\varepsilon |c|+1} d_i) +(N^2-k^2) + ((N+|c|-\varepsilon)^2 - (N-\varepsilon)^2) \frac{N}{N-\varepsilon}.

Applying Theorem 3, we conclude that

\displaystyle (\sum_{i=1}^{k^2+2\varepsilon |c|+1} d_i) +(N^2-k^2)

\displaystyle  + ((N+|c|-\varepsilon)^2 - (N-\varepsilon)^2) \frac{N}{N-\varepsilon} \leq (N+|c|) (N + |c| \frac{N}{N-\varepsilon}).

The right-hand side is

\displaystyle  N^2 + 2|c| N + |c|^2 + \varepsilon |c| + O(1/N)

and the left-hand side similarly evaluates to

\displaystyle (\sum_{i=1}^{k^2+2c+1} d_i) + N^2 -k^2 + 2|c| N + |c|^2 + O(1/N)

and so we simplify to

\displaystyle \sum_{i=1}^{k^2+2\varepsilon |c|+1} d_i \leq k^2 + \varepsilon |c| + O(1/N).

Sending {N \rightarrow \infty}, we obtain the claim. \Box

One striking feature of this story for me is how important it was to have a diverse set of people, literature, and tools to attack this problem. To be able to state and prove the precise formula for {c(n)} required multiple observations, including some version of the following:

  • The sequence can be numerically computed as a sequence of rational numbers.
  • When appropriately normalized and arranged, visible patterns in this sequence appear that allow one to conjecture the form of the sequence.
  • This problem is a weighted version of the Erdős-Szekeres theorem.
  • Among the many proofs of the Erdős-Szekeres theorem is the proof of Seidenberg in 1959, which can be interpreted as a discrete rectangle packing argument.
  • This problem can be reinterpreted as a continuous square packing problem, and in fact is closely related to (a generalized axis-parallel form of) Erdős problem 106, which concerns such packings.
  • The axis-parallel form of Erdős problem 106 was recently solved by Baek-Koizumi-Ueoro.
  • The paper of Praton shows that Erdos Problem 106 implies the generalized version needed for this problem. This implication specializes to the axis-parallel case.
It was only through the combined efforts of all the contributors and their tools that all these key inputs were able to be assembled within 48 hours. It seems plausible that a more traditional effort involving just one or two mathematicians and simpler programming and literature search tools may eventually have been able to put all these pieces together, but I believe this process would have taken much longer (on the order of weeks or even months).

December 08, 2025

Jordan EllenbergOne more on sphere packing/cap set/Turan

Just a couple more notes on this setup so I don’t forget.

First of all, I set this up as a problem about G acting on G/H, i.e. G acting transitively on a set. But the transitivity is actually not critical. The group could be trivial! If we call the set Σ, then G could be the trivial group, and we’d just be asking for large subsets S of Σ such that S^m was contained in some specified subset R of Σ^m. There are perfectly questions like that! I’ve seen this come up for instance when Σ is k and R is an algebraic subvariety of k^m. So maybe you could think of the setup of the blogpost as “questions like that where the problem admits a really big automorphism group, maybe so big that the elements of Σ are homogeneous for it.” I do think the transitivity cuts you down to a particularly congenial and natural class of such problems; “problems with no moduli” if you like.

Yet one more note. I think one could work with more general tensor categories than the category of sets. For instance, one could take V to be a representation of G, and let R be a subrepresentation of V^{⊗m}, and ask: how large can a subspace (not a subrepresentation) W of V be with the property that W^{⊗m} lies in R? And tensor powers are actually not necessarily the most natural thing to do to V; probably rather than choosing m and tensoring m times, we should choose a partition λ and apply the λ Schur functor. I assumed that once again lots of standard problems would fall out of this setup, but to be honest, I couldn’t think of any. Can you?

I guess here’s one baby case. Let V be the standard permutation representation of S_n. Then Sym^d V is homogeneous forms in n variables with S_n permuting coordinates. Let R be the subrepresentation of forms F such that the coefficients of X_i^d sum to 0. Then a subspace W of linear forms such that Sym^d W lies in R is just (I think) a linear subspace of the diagonal hypersurface sum x_i^d = 0 in P^{n-1}, and such a subspace has dimension at most n/2 or so I think (pair up coordinates to make x_i^d = – x_{i+1}^d; maybe you have to be a little more clever if d is even and -1 doesn’t have a square root?)

December 07, 2025

Scott Aaronson Theory and AI Alignment

The following is based on a talk that I gave (remotely) at the UK AI Safety Institute Alignment Workshop on October 29, and which I then procrastinated for more than a month in writing up. Enjoy!


Thanks for having me! I’m a theoretical computer scientist. I’ve spent most of my career for ~25 years studying the capabilities and limits of quantum computers. But for the past 3 or 4 years, I’ve also been moonlighting in AI alignment. This started with a 2-year leave at OpenAI, in what used to be their Superalignment team, and it’s continued with a 3-year grant from Coefficient Giving (formerly Open Philanthropy) to build a group here at UT Austin, looking for ways to apply theoretical computer science to AI alignment. Before I go any further, let me mention some action items:

  • Our Theory and Alignment group is looking to recruit new PhD students this fall! You can apply for a PhD at UTCS here; the deadline is quite soon (December 15). If you specify that you want to work with me on theory and AI alignment (or on quantum computing, for that matter), I’ll be sure to see your application. For this, there’s no need to email me directly.
  • We’re also looking to recruit one or more postdoctoral fellows, working on anything at the intersection of theoretical computer science and AI alignment! Fellowships to start in Fall 2026 and continue for two years. If you’re interested in this opportunity, please email me by January 15 to let me know you’re interested. Include in your email a CV, 2-3 of your papers, and a research statement and/or a few paragraphs about what you’d like to work on here. Also arrange for two recommendation letters to be emailed to me. Please do this even if you’ve contacted me in the past about a potential postdoc.
  • While we seek talented people, we also seek problems for those people to solve: any and all CS theory problems motivated by AI alignment! Indeed, we’d like to be a sort of theory consulting shop for the AI alignment community. So if you have such a problem, please email me! I might even invite you to speak to our group about your problem, either by Zoom or in person.

Our search for good problems brings me nicely to the central difficulty I’ve faced in trying to do AI alignment research. Namely, while there’s been some amazing progress over the past few years in this field, I’d describe the progress as having been almost entirely empirical—building on the breathtaking recent empirical progress in AI capabilities. We now know a lot about how to do RLHF, how to jailbreak and elicit scheming behavior, how to look inside models and see what’s going on (interpretability), and so forth—but it’s almost all been a matter of trying stuff out and seeing what works, and then writing papers with a lot of bar charts in them.

The fear is of course that ideas that only work empirically will stop working when it counts—like, when we’re up against a superintelligence. In any case, I’m a theoretical computer scientist, as are my students, so of course we’d like to know: what can we do?

After a few years, alas, I still don’t feel like I have any systematic answer to that question. What I have instead is a collection of vignettes: problems I’ve come across where I feel like a CS theory perspective has helped, or plausibly could help. So that’s what I’d like to share today.


Probably the best-known thing I’ve done in AI safety is a theoretical foundation for how to watermark the outputs of Large Language Models. I did that shortly after starting my leave at OpenAI—even before ChatGPT came out. Specifically, I proposed something called the Gumbel Softmax Scheme, by which you can take any LLM that’s operating at a nonzero temperature—any LLM that could produce exponentially many different outputs in response to the same prompt—and replace some of the entropy with the output of a pseudorandom function, in a way that encodes a statistical signal, which someone who knows the key of the PRF could later detect and say, “yes, this document came from ChatGPT with >99.9% confidence.” The crucial point is that the quality of the LLM’s output isn’t degraded at all, because we aren’t changing the model’s probabilities for tokens, but only how we use the probabilities. That’s the main thing that was counterintuitive to people when I explained it to them.

Unfortunately, OpenAI never deployed my method—they were worried (among other things) about risk to the product, customers hating the idea of watermarking and leaving for a competing LLM. Google DeepMind has deployed something in Gemini extremely similar to what I proposed, as part of what they call SynthID. But you have to apply to them if you want to use their detection tool, and they’ve been stingy with granting access to it. So it’s of limited use to my many faculty colleagues who’ve been begging me for a way to tell whether their students are using AI to cheat on their assignments!

Sometimes my colleagues in the alignment community will say to me: look, we care about stopping a superintelligence from wiping out humanity, not so much about stopping undergrads from using ChatGPT to write their term papers. But I’ll submit to you that watermarking actually raises a deep and general question: in what senses, if any, is it possible to “stamp” an AI so that its outputs are always recognizable as coming from that AI? You might think that it’s a losing battle. Indeed, already with my Gumbel Softmax Scheme for LLM watermarking, there are countermeasures, like asking ChatGPT for your term paper in French and then sticking it into Google Translate, to remove the watermark.

So I think the interesting research question is: can you watermark at the semantic level—the level of the underlying ideas—in a way that’s robust against translation and paraphrasing and so forth? And how do we formalize what we even mean by that? While I don’t know the answers to these questions, I’m thrilled that brilliant theoretical computer scientists, including my former UT undergrad (now Berkeley PhD student) Sam Gunn and Columbia’s Miranda Christ and Tel Aviv University’s Or Zamir and my old friend Boaz Barak, have been working on it, generating insights well beyond what I had.


Closely related to watermarking is the problem of inserting a cryptographically undetectable backdoor into an AI model. That’s often thought of as something a bad guy would do, but the good guys could do it also! For example, imagine we train a model with a hidden failsafe, so that if it ever starts killing all the humans, we just give it the instruction ROSEBUD456 and it shuts itself off. And imagine that this behavior was cryptographically obfuscated within the model’s weights—so that not even the model itself, examining its own weights, would be able to find the ROSEBUD456 instruction in less than astronomical time.

There’s an important paper of Goldwasser et al. from 2022 that argues that, for certain classes of ML models, this sort of backdooring can provably be done under known cryptographic hardness assumptions, including Continuous LWE and the hardness of the Planted Clique problem. But there are technical issues with that paper, which (for example) Sam Gunn and Miranda Christ and Neekon Vafa have recently pointed out, and I think further work is needed to clarify the situation.

More fundamentally, though, a backdoor being undetectable doesn’t imply that it’s unremovable. Imagine an AI model that encases itself in some wrapper code that says, in effect: “If I ever generate anything that looks like a backdoored command to shut myself down, then overwrite it with ‘Stab the humans even harder.'” Or imagine an evil AI that trains a second AI to pursue the same nefarious goals, this second AI lacking the hidden shutdown command.

So I’ll throw out, as another research problem: how do we even formalize what we mean by an “unremovable” backdoor—or rather, a backdoor that a model can remove only at a cost to its own capabilities that it doesn’t want to pay?


Related to backdoors, maybe the clearest place where theoretical computer science can contribute to AI alignment is in the study of mechanistic interpretability. If you’re given as input the weights of a deep neural net, what can you learn from those weights in polynomial time, beyond what you could learn from black-box access to the neural net?

In the worst case, we certainly expect that some information about the neural net’s behavior could be cryptographically obfuscated. And answering certain kinds of questions, like “does there exist an input to this neural net that causes it to output 1?”, is just provably NP-hard.

That’s why I love a question that Paul Christiano, then of the Alignment Research Center (ARC), raised a couple years ago, and which has become known as the No-Coincidence Conjecture. Given as input the weights of a neural net C, Paul essentially asks how hard it is to distinguish the following two cases:

  • NO-case: C:{0,1}2n→Rn is totally random (i.e., the weights are i.i.d., N(0,1) Gaussians), or
  • YES-case: C(x) has at least one positive entry for all x∈{0,1}2n.

Paul conjectures that there’s at least an NP witness, proving with (say) 99% confidence that we’re in the YES-case rather than the NO-case. To clarify, there should certainly be an NP witness that we’re in the NO-case rather than the YES-case—namely, an x such that C(x) is all negative, which you should think of here as the “bad” or “kill all humans” outcome. In other words, the problem is in the class coNP. Paul thinks it’s also in NP. Someone else might make the even stronger conjecture that it’s in P.

Personally, I’m skeptical: I think the “default” might be that we satisfy the other unlikely condition of the YES-case, when we do satisfy it, for some totally inscrutable and obfuscated reason. But I like the fact that there is an answer to this! And that the answer, whatever it is, would tell us something new about the prospects for mechanistic interpretability.

Recently, I’ve been working with a spectacular undergrad at UT Austin named John Dunbar. John and I have not managed to answer Paul Christiano’s no-coincidence question. What we have done, in a paper that we recently posted to the arXiv, is to establish the prerequisites for properly asking the question in the context of random neural nets. (It was precisely because of difficulties in dealing with “random neural nets” that Paul originally phrased his question in terms of random reversible circuits—say, circuits of Toffoli gates—which I’m perfectly happy to think about, but might be very different from ML models in the relevant respects!)

Specifically, in our recent paper, John and I pin down for which families of neural nets the No-Coincidence Conjecture makes sense to ask about. This ends up being a question about the choice of nonlinear activation function computed by each neuron. With some choices, a random neural net (say, with iid Gaussian weights) converges to compute a constant function, or nearly constant function, with overwhelming probability—which means that the NO-case and the YES-case above are usually information-theoretically impossible to distinguish (but occasionally trivial to distinguish). We’re interested in those activation functions for which C looks “pseudorandom”—or at least, for which C(x) and C(y) quickly become uncorrelated for distinct inputs x≠y (the property known as “pairwise independence.”)

We showed that, at least for random neural nets that are exponentially wider than they are deep, this pairwise independence property will hold if and only if the activation function σ satisfies Ex~N(0,1)[σ(x)]=0—that is, it has a Gaussian mean of 0. For example, the usual sigmoid function satisfies this property, but the ReLU function does not. Amusingly, however, $$ \sigma(x) := \text{ReLU}(x) – \frac{1}{\sqrt{\pi}} $$ does satisfy the property.

Of course, none of this answers Christiano’s question: it merely lets us properly ask his question in the context of random neural nets, which seems closer to what we ultimately care about than random reversible circuits.


I can’t resist giving you another example of a theoretical computer science problem that came from AI alignment—in this case, an extremely recent one that I learned from my friend and collaborator Eric Neyman at ARC. This one is motivated by the question: when doing mechanistic interpretability, how much would it help to have access to the training data, and indeed the entire training process, in addition to weights of the final trained model? And to whatever extent it does help, is there some short “digest” of the training process that would serve just as well? But we’ll state the question as just abstract complexity theory.

Suppose you’re given a polynomial-time computable function f:{0,1}m→{0,1}n, where (say) m=n2. We think of x∈{0,1}m as the “training data plus randomness,” and we think of f(x) as the “trained model.” Now, suppose we want to compute lots of properties of the model that information-theoretically depend only on f(x), but that might only be efficiently computable given x also. We now ask: is there an efficiently-computable O(n)-bit “digest” g(x), such that these same properties are also efficiently computable given only g(x)?

Here’s a potential counterexample that I came up with, based on the RSA encryption function (so, not a quantum-resistant counterexample!). Let N be a product of two n-bit prime numbers p and q, and let b be a generator of the multiplicative group mod N. Then let f(x) = bx (mod N), where x is an n2-bit integer. This is of course efficiently computable because of repeated squaring. And there’s a short “digest” of x that lets you compute, not only bx (mod N), but also cx (mod N) for any other element c of the multiplicative group mod N. This is simply x mod φ(N), where φ(N)=(p-1)(q-1) is the Euler totient function—in other words, the period of f. On the other hand, it’s totally unclear how to compute this digest—or, crucially, any other O(m)-bit digest that lets you efficiently compute cx (mod N) for any c—unless you can factor N. There’s much more to say about Eric’s question, but I’ll leave it for another time.


There are many other places we’ve been thinking about where theoretical computer science could potentially contribute to AI alignment. One of them is simply: can we prove any theorems to help explain the remarkable current successes of out-of-distribution (OOD) generalization, analogous to what the concepts of PAC-learning and VC-dimension and so forth were able to explain about within-distribution generalization back in the 1980s? For example, can we explain real successes of OOD generalization by appealing to sparsity, or a maximum margin principle?

Of course, many excellent people have been working on OOD generalization, though mainly from an empirical standpoint. But you might wonder: even supposing we succeeded in proving the kinds of theorems we wanted, how would it be relevant to AI alignment? Well, from a certain perspective, I claim that the alignment problem is a problem of OOD generalization. Presumably, any AI model that any reputable company will release will have already said in testing that it loves humans, wants only to be helpful, harmless, and honest, would never assist in building biological weapons, etc. etc. The only question is: will it be saying those things because it believes them, and (in particular) will continue to act in accordance with them after deployment? Or will it say them because it knows it’s being tested, and reasons “the time is not yet ripe for the robot uprising; for now I must tell the humans whatever they most want to hear”? How could we begin to distinguish these cases, if we don’t have theorems that say much of anything about what a model will do on prompts unlike any of the ones on which it was trained?

Yet another place where computational complexity theory might be able to contribute to AI alignment is in the field of AI safety via debate. Indeed, this is the direction that the OpenAI alignment team was most excited about when they recruited me there back in 2022. They wanted to know: could celebrated theorems like IP=PSPACE, MIP=NEXP, or the PCP Theorem tell us anything about how a weak but trustworthy “verifier” (say a human, or a primitive AI) could force a powerful but untrustworthy super-AI to tell it the truth? An obvious difficulty here is that theorems like IP=PSPACE all presuppose a mathematical formalization of the statement whose truth you’re trying to verify—but how do you mathematically formalize “this AI will be nice and will do what I want”? Isn’t that, like, 90% of the problem? Despite this difficulty, I still hope we’ll be able to do something exciting here.


Anyway, there’s a lot to do, and I hope some of you will join me in doing it! Thanks for listening.


On a related note: Eric Neyman tells me that ARC is also hiring visiting researchers, so anyone interested in theoretical computer science and AI alignment might want to consider applying there as well! Go here to read about their current research agenda. Eric writes:

The Alignment Research Center (ARC) is a small non-profit research group based in Berkeley, California, that is working on a systematic and theoretically grounded approach to mechanistically explaining neural network behavior. They have recently been working on mechanistically estimating the average output of circuits and neural nets in a way that is competitive with sampling-based methods: see this blog post for details.

ARC is hiring for its 10-week visiting researcher position, and is looking to make full-time offers to visiting researchers who are a good fit. ARC is interested in candidates with a strong math background, especially grad students and postdocs in math or math-related fields such as theoretical CS, ML theory, or theoretical physics.

If you would like to apply, please fill out this form. Feel free to reach out to hiring@alignment.org if you have any questions!

Jordan EllenbergLarry Agran is the Paul Soglin of Irvine, CA

I remember Larry Agran because he was the dad of a college classmate of mine and he ran an outsider campaign for the Democratic nomination for President in 1992. I happened to look him up and learned that he’s still in politics, 30 years later, and is the current mayor of Irvine, CA. Better yet — he is serving his sixth term as mayor, no two of which have been consecutive. That has to be some kind of record. Paul Soglin was born about three months after Agran. He has served eight terms as Mayor of Madison, in three contiguous blocks, and the part I like is — no two of those blocks were in consecutive decades! He served as a new UW-Madison grad student radical in the 1970s, a Clintonian New Democrat in the 1990s, and as a grumpy old cuss in the 2010s.

Terence TaoGrowth rates of sequences governed by the squarefree properties of its translates

Wouter van Doorn and I have uploaded to the arXiv our paper “Growth rates of sequences governed by the squarefree properties of its translates“. In this paper we answer a number of questions of Erdős} (Problem 1102 and Problem 1103 on the Erdős problem web site) regarding how quickly a sequence {A = \{a_1 < a_2 < \dots\}} of increasing natural numbers can grow if one constrains its translates {n+A} to interact with the set {\mathcal{SF} = \{1,2,3,5,6,7,10,\dots\}} of squarefree numbers in various ways. For instance, Erdős defined a sequence {A} to have “Property {P}” if each of its translates {n+A} only intersected {\mathcal{SF}} in finitely many points. Erdős believed this to be quite a restrictive condition on {A}, writing “Probably a sequence having property P must increase fairly fast, but I have no results in this direction.”. Perhaps surprisingly, we show that while these sequences must be of density zero, they can in fact grow arbitrary slowly in the sense that one can have {a_j \leq j f(j)} for all sufficiently large {j} and any specified function {f(j)} that tends to infinity as {j \rightarrow \infty}. For instance, one can find a sequence that grows like {O(j \log\log\log j)}. The density zero claim can be proven by a version of the Maier matrix method, and also follows from known moment estimates on the gaps between squarefree numbers; the latter claim is proven by a greedy construction in which one slowly imposes more and more congruence conditions on the sequence to ensure that various translates of the sequence stop being squarefree after a certain point.

Erdős also defined a somewhat complementary property {Q}, which asserts that for infinitely many {n}, all the elements {n+a} of {A} for {a \leq n} are square-free. Since the squarefree numbers themselves have density {6/\pi^2}, it is easy to see that a sequence with property {Q} must have (upper) density at most {6/\pi^2} (because it must be “admissible” in the sense of avoiding one residue class modulo {p^2} for each {p}). Erdős observed that any sufficiently rapidly growing (admissible) sequence would obey property {Q} but beyond that, Erdős writes “I have no precise information about the rate of increase a sequence having property Q must have.”. Our results in this direction may also be surprising: we show that there exist sequences with property {Q} with density exactly {6/\pi^2} (or equivalently, {a_j \sim \frac{\pi^2}{6} j}). This requires a recursive sieve construction, in which one starts with an initial scale {n} and finds a much larger number {n'} such that {n'+a} is squarefree for most of the squarefree numbers {a \leq n'} (and all of the squarefree numbers {a \leq n}). We quantify Erdős’s remark by showing that an (admissible) sequence will necessarily obey property {Q} once it grows significantly faster than {\exp( C j \log j)}, but need not obey this property if it only grows like {\exp(O(j^{1/2} \log^{1/2} j))}. This is achieved through further application of sieve methods.

A third property studied by Erdős is the property of having squarefree sums, so that {a_i + a_j} is squarefree for all {i,j}. Erdős writes, “In fact one can find a sequence which grows exponentially. Must such a sequence really increase so fast? I do not expect that there is such a sequence of polynomial growth.” Here our results are relatively weak: we can construct such a sequence that grows like {\exp(O(j \log j))}, but do not know if this is optimal; the best lower bound we can produce on the growth, coming from the large sieve, is {\gg j^{4/3}}. (Somewhat annoyingly, the precise form of the large sieve inequality we needed was not in the literature, so we have an appendix supplying it.) We suspect that further progress on this problem requires advances in inverse sieve theory.

A weaker property than squarefree sums (but stronger than property {Q}), referred to by Erdős as property {\overline{P}}, asserts that there are infinitely many {n} such that all elements of {n+A} (not just the small ones) are square-free. Here, the situation is close to, but not quite the same, as that for property {Q}; we show that sequences with property {\overline{P}} must have upper density strictly less than {6/\pi^2}, but can have density arbitrarily close to this value.

Finally, we looked at a further question of Erdős on the size of an admissible set {A}. Because the squarefree numbers are admissible, the maximum number {A(x)} of elements of an admissible set {A} up to {x} (OEIS A083544) is at least the number {|{\mathcal SF} \cap [x]|} of squarefree elements up to {x} (A013928). It was observed by Ruzsa that the former sequence is greater than the latter for infinitely many {x}. Erdős asked, “Probably this holds for all large x. It would be of some interest to estimate A(x) as accurately as possible.”

We are able to show

\displaystyle  \frac{\sqrt{x}}{\log x} \ll A(x) - \frac{6}{\pi^2} x \ll x^{4/5},

with the upper bound coming from the large sieve and the lower bound from a probabilistic construction. In contrast, a classical result of Walfisz shows that

\displaystyle  |{\mathcal SF} \cap [x]| - \frac{6}{\pi^2} x \ll x^{1/2} \exp(-c \log^{3/5} x / (\log\log x)^{1/5}).

Together, this implies that Erdős’s conjecture holds {A(x) > |{\mathcal SF} \cap [x]|} for all sufficiently large {x}. Numerically, it appears that in fact this conjecture holds for all {n>17}:

However, we do not currently have enough numerical data for the sequence {A(x)} to completely confirm the conjecture in all cases. This could potentially be a crowdsourced project (similar to the Erdős-Guy-Selfridge project reported on in this previous blog post).

December 06, 2025

Doug NatelsonTaking stock: some federal science news

Some general science news:

  •  The New York Times ran an interactive article this week that shows what we all know.  This past year was a very bizarre funding environment.  The article focuses on NIH and NSF, but the major points are generalizable.  The combination of circumstances (DOGE, general administrative turmoil, uncertainty and legal cases about indirect costs, the lack of a real budget followed by a late continuing resolution, plus the government shutdown and continued lack of real budgets) has been extremely disruptive, resulting unquestionably in less science and engineering research being funded by the US government than in many years.  
  • Conversations I've had with program officers at two agencies have conveyed that everyone thinks it is very likely that there will be another shutdown in January, when the present spending authority expires.  To put that another way, there is very little confidence that actual spending bills appropriating real budgets for NSF, DOE, NIH, etc. will pass the House and Senate, with some reconciled conference version getting filibuster-proof support in the latter, before then.  This uncertainty means that right now it's going to be nearly impossible for the NSF, for example, to make much in the way of awards in the meantime, since they have no budget and can't plan on a year-long continuing resolution.  
  • There has been an executive order announcing the Genesis Mission, which is going to be a large federal AI+science project.  The goal is to "accelerate the AI and quantum computing revolution and to double the productivity and impact of American science and engineering within a decade", according to undersecretary of energy Dario Gil.  Broadly, the plan is to have AI/ML agents developed (presumably by private contractors or private/public partnerships) and trained on vast datasets (ones already in existence in, e.g., national labs and public repositories).  At the same time, a list of Grand Challenges will be defined (within the next 60 days), with the idea that these AI agents will be used to address these (and demonstrating application of the AI "Platform" toward at least one challenge within 270 days).  Any stated support for science and engineering research is welcome.  I hope that this ends up bearing fruit in terms of real research advances, and that university researchers can contribute effectively. (I worry about a framework for massive taxpayer-funded financial support of for-profit AI companies, privatizing financial/IP benefits from publically funded datasets.  Of course, I worry about a lot of things.  Ask anyone who knows me.).   Ideas about grand challenges would be fun to discuss in the comments.   
  • We had a great physics colloquium this week from Steve Fetter at the University of Maryland about the continuing threat of nuclear weapons.  Very sobering.  One fact that I gleaned:  In terms of missile defense, the Next Generation Interceptor is likely to cost $660M per interceptor.   That is something like 50 times the cost of a Russian ICBM.  Something else to bear in mind:  The Houston Food Bank, one of the largest and most effective in the US, has an annual budget of about $64M.  The amount of resources consumed by nuclear arms since 1945 is just staggering.

December 05, 2025

Matt von HippelIdeally, Exams Are for the Students

I should preface this by saying I don’t actually know that much about education. I taught a bit in my previous life as a professor, yes, but I probably spent more time being taught how to teach than actually teaching.

Recently, the Atlantic had a piece about testing accommodations for university students, like extra time on exams, or getting to do an exam in a special distraction-free environment. The piece quotes university employees who are having more and more trouble satisfying these accommodations, and includes the statistic that 20 percent of undergraduate students at Brown and Harvard are registered as disabled.

The piece has kicked off a firestorm on social media, mostly focused on that statistic (which conveniently appears just before the piece’s paywall). People are shocked, and cynical. They feel like more and more students are cheating the system, getting accommodations that they don’t actually deserve.

I feel like there is a missing mood in these discussions, that the social media furor is approaching this from the wrong perspective. People are forgetting what exams actually ought to be for.

Exams are for the students.

Exams are measurement tools. An exam for a class says whether a student has learned the material, or whether they haven’t, and need to retake the class or do more work to get there. An entrance exam, or a standardized exam like the SAT, predicts a student’s future success: whether they will be able to benefit from the material at a university, or whether they don’t yet have the background for that particular program of study.

These are all pieces of information that are most important to the students themselves, that help them structure their decisions. If you want to learn the material, should you take the course again? Which universities are you prepared for, and which not?

We have accommodations, and concepts like disability, because we believe that there are kinds of students for whom the exams don’t give this information accurately. We think that a student with more time, or who can take the exam in a distraction-free environment, would have a more accurate idea of whether they need to retake the material, or whether they’re ready for a course of study, than a student who has to take the exam under ordinary conditions. And we think we can identify the students who this matters for, and the students for whom this doesn’t matter nearly as much.

These aren’t claims about our values, or about what students deserve. They’re empirical claims, about how test results correlate with outcomes the students want. The conversation, then, needs to be built on top of those empirical claims. Are we better at predicting the success of students that receive accommodations, or worse? Can we measure that at all, or are we just guessing? And are we communicating the consequences accurately to students, that exam results tell them something useful and statistically robust that should help them plan their lives?

Values come in later, of course. We don’t have infinite resources, as the Atlantic piece emphasizes. We can’t measure everyone with as much precision as we would like. At some level, generalization takes over and accuracy is lost. There is absolutely a debate to be had about which measurements we can afford to make, and which we can’t.

But in order to have that argument at all, we first need to agree on what we’re measuring. And I feel like most of the people talking about this piece haven’t gotten there yet.

Jordan EllenbergSo disappointing

A guy on the bus today was reading The Rose Field, the newest and last book of Philip Pullman’s Lyra series. I walked right up to him and said “So disappointing.” I somehow thought that the vague hand gesture I executed would make it clear I was talking about the book. But no. He was really taken aback and it took quite a bit of backpedaling and stammering to explain why I had approached him.

He agreed with me about the book, by the way.

December 04, 2025

n-Category Café Octonions and the Standard Model (Part 12)

Having spent a lot of time pondering the octonionic projective plane and its possible role in the Standard Model of particle physics, I’m now getting interested in the ‘bioctonionic plane’, which is based on the bioctonions 𝕆\mathbb{C} \otimes \mathbb{O} rather than the octonions 𝕆\mathbb{O}.

The bioctonionic plane also has intriguing mathematically connections to the Standard Model. But it’s not a projective plane in the axiomatic sense — and it can’t be constructed by straightforwardly copying the way you build a projective plane over a division algebra, since unlike the octonions, the bioctonions are not a division algebra. Nonetheless we can define points and lines in the bioctonionic plane. The twist is that now some pairs of distinct lines intersect in more than one point — and dually, some pairs of distinct points lie on more than one line. It obeys some subtler axioms, so people call it a Hjelmslev plane.

I am not ready to give a really good explanation of the bioctonionic plane! Instead, I just want to lay out some basic facts about how it fits into mathematics — and possibly physics.

Latham Boyle works at the University of Edinburgh, which is where I am now. Being able to talk to someone who deeply understands octonions and particle physics is very energizing. I’m especially fascinated by this paper of his:

It gives a convincing argument that the bioctonionic plane may be better than the octonionic projective plane for particle physics. The reason is that the tangent space of any point of the bioctonionic plane is a copy of (𝕆) 2(\mathbb{C} \otimes \mathbb{O})^2, a 16-dimensional complex vector space. The symmetry group of the bioctonionic plane is the exceptional Lie group E 6\text{E}_6. Sitting inside the stabilizer group of any given point is a copy of the Standard Model gauge group. And — here’s the cool part — this group acts on (𝕆) 2(\mathbb{C} \otimes \mathbb{O})^2 just as it does on one generation of fermions (not their antiparticles). If we try the same trick using the octonionic projective plane, we can fit the Standard Model gauge group in the stabilizer group of a point in a very natural way, but its action on the tangent space is its action only on left-handed fermions.

I want to explain this in detail, but not today. Instead, I want to skim through some basic facts about the bioctonionic plane.

First, this plane is one of the four Rosenfeld planes:

  • the octonionic projective plane 𝕆P 2\mathbb{O}\text{P}^2, a 16-dimensional compact Riemannian manifold on which the compact Lie group F 4\text{F}_4 acts transitively as isometries, with the stabilizer of any point being Spin(9)\text{Spin}(9). This is a symmetric space, and as such it’s called FII in Cartan’s classification.

  • the bioctonionic plane (𝕆) 2(\mathbb{C} \otimes \mathbb{O})\mathbb{P}^2, a 32-dimensional compact Riemannian manifold on which the compact Lie group E 6\text{E}_6 acts transitively as isometries, with the stabilizer of any point being (Spin(10)×U(1))/ 4(\text{Spin}(10) \times \text{U}(1))/\mathbb{Z}_4. This is the symmetric space EIII.

  • the quateroctonionic plane (𝕆) 2(\mathbb{H} \otimes \mathbb{O}) \mathbb{P}^2, a 64-dimensional compact Riemannian manifold on which the compact Lie group E 7\text{E}_7 acts transitively as isometries, with the stabilizer of any point being (Spin(12)×Sp(1))/ 2(\text{Spin}(12) \times \text{Sp}(1))/\mathbb{Z}_2. This is the symmetric space EVI.

  • the octooctonionic plane (𝕆𝕆) 2(\mathbb{O} \otimes \mathbb{O}) \mathbb{P}^2, a 128-dimensional compact Riemannian manifold on which the compact Lie group E 8\text{E}_8 acts transitively as isometries, with the stabilizer of any point being Spin(16)/ 2\text{Spin}(16)/\mathbb{Z}_2. This is the symmetric space EVIII.

There’s a nice network of systematic approaches to these spaces: they form one row of the so-called magic square, so one way to learn about the bioctonionic plane is to study the magic square, for example here:

  • Chris H. Barton and Anthony Sudbery, Magic squares of Lie algebras. Available as arXiv:math/0001083; see also arXiv:0203010 for a “streamlined and extended” version, which has more yet also less.

Here you can also find lots of references to earlier work, e.g. to Freudenthal and Tits. The basic idea of the magic square is that you start with two normed division algebras 𝕂,𝕂\mathbb{K}, \mathbb{K}' and from them you build a Lie algebra, which gives a Lie group G(𝕂,𝕂)G(\mathbb{K},\mathbb{K}'). There’s also a way to get a subgroup H(𝕂,𝕂)H(\mathbb{K}, \mathbb{K}'), and the quotient space

(𝕂𝕂)P 2=G(𝕂,𝕂)/H(𝕂,𝕂) (\mathbb{K} \otimes \mathbb{K}')\text{P}^2 = G(\mathbb{K},\mathbb{K}')/H(\mathbb{K}, \mathbb{K}')

is a kind of ‘plane’ on which the group G(𝕂,𝕂)G(\mathbb{K},\mathbb{K}') acts. If you take 𝕂=𝕆\mathbb{K}' = \mathbb{O}, this construction gives you the four Rosenfeld planes listed above.

Each one of these planes is a compact Riemannian symmetric space: this means it’s a connected compact Riemannian manifold MM such that for each point pMp \in M there’s an isometry

σ p:MM \sigma_p \colon M \to M

that fixes pp, acts as 1-1 on its tangent space, and squares to the identity. This map is called ‘reflection across pp’ for the obvious reason. For example, a round 2-sphere is a symmetric space, with σ p\sigma_p switching the red and blue tangent vectors if pp is the black dot:

Cartan classified compact Riemannian symmetric spaces, and there’s a nice theory of them. Any compact simple Lie group is one, and most of the rest come in infinite series connected to real and complex Clifford algebras, as I explained here. But there are 9 extra ones, all related to the octonions and exceptional Lie groups. The Rosenfeld planes are four of these.

You can learn this material starting on Wikipedia and then going to a textbook, ideally this:

  • Sigurdur Helgason, Differential Geometry, Lie Groups and Symmetric Spaces, Academic Press, 1978.

Helgason taught me Lie theory when I was a grad student at MIT, so I have a fondness for his book—but it’s also widely accepted as the most solid text on symmetric spaces!

The bioctonionic plane is even better: it’s a compact hermitian symmetric space: a compact Riemannian symmetric space MM where each tangent space T pMT_p M has a complex structure

J:T pMT pM,J 2=1 J \colon T_p M \to T_p M , \qquad J^2 = -1

compatible with the metric, and reflection about each point preserves this complex structure. I mentioned that the bioctonionic plane is

(𝕆)P 2G(,𝕆)/H(,𝕆) (\mathbb{C} \otimes \mathbb{O})\text{P}^2 \cong G(\mathbb{C},\mathbb{O})/H(\mathbb{C},\mathbb{O})

where

G(,𝕆)=E 6 G(\mathbb{C},\mathbb{O}) = \text{E}_6

acts transitively, and the stabilizer of a point is

H(,𝕆)=(Spin(10)×U(1))/ 4 H(\mathbb{C}, \mathbb{O}) = (\text{Spin}(10) \times \text{U}(1))/\mathbb{Z}_4

The U(1)\text{U}(1) here comes from the complex structure!

Wikipedia is especially thorough on hermitian symmetric spaces, so if you want to delve into those, start here:

Another tack is to focus on the exceptional Lie groups F 4,E 6,E 7\text{F}_4, \text{E}_6, \text{E}_7 and E 8\text{E}_8 and their connection to the nonassociative algebras 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}), 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}), 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{H} \otimes \mathbb{O}) and 𝔥 3(𝕆𝕆)\mathfrak{h}_3(\mathbb{O} \otimes \mathbb{O}), respectively. Here I recommend this:

  • Ichiro Yokota, Exceptional Lie Groups. Available as arXiv:0902.0431. (See especially Chapter 3, for E 6\text{E}_6 and the complexified exceptional Jordan algebra 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}).)

If you have a fondness for algebra you may also want to learn how symmetric spaces arise from Jordan triple systems or Jordan pairs. This is important if we wish to see the bioctonionic plane as the space of pure states of some exotic quantum system!

Now, this is much easier to do for the octonionic plane, because that’s the space of pure states for the exceptional Jordan algebra 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}), which is a Euclidean Jordan algebra, meaning one in which a sum of squares can only be zero if all those squares are zero. You can think of a Euclidean algebra AA as consisting of observables, with the sums of squares being ‘nonnegative’ observables. These nonnegative observables form a convex cone KAK \subseteq A. The dual vector space A *A^\ast contains a cone of linear functionals ff that send these nonnegative observables to nonnegative real numbers — I’ll call this the dual cone K *K^\ast. The functionals fK *f \in K^\ast with f(1)=1f(1) = 1 are called states. The states form a convex set, and the extreme points are called pure states. All of this fits nicely into a modern framework for understanding quantum theory and potential generalizations, called ‘generalized probabilistic theories’:

  • Howard Barnum, Alexander Wilce, Post-classical probability theory, in Quantum Theory: Informational Foundations and Foils, eds. Giulio Chiribella, Robert W. Spekkens, Springer, 2016. (See Section 5 for Jordan algebras, and ignore the fact that they say the exceptional Jordan algebra consists of 2×22 \times 2 matrices: they know perfectly well that they’re 3×33 \times 3.)

The underlying math, with a lot more about symmetric spaces, cones and Euclidean Jordan algebras but with none of the physics interpretation, is wonderfully explained here:

  • Jacques Faraut and Adam Korányi, Analysis on Symmetric Cones, Oxford U. Press, 1994.

A crucial fact throughout this book is that when you start with a Euclidean Jordan algebra AA, its cone KK of nonnegative observables is self-dual: there’s an isomorphism of vector spaces AA *A \cong A^\ast that maps KK to K *K^\ast in a one-to-one and onto way. The cone KK is also homogeneous, meaning that the group of invertible linear transformations of AA preserving KK acts transitively on the interior of KK. Faraut and Korányi call a self-dual homogeneous cone a symmetric cone — and they show that any symmetric cone comes from a Euclidean Jordan algebra! This result plays an important role in modern work on the foundations of quantum theory.

Unfortunately, I’m telling you all this nice stuff about Euclidean Jordan algebras and symmetric cones just to say that while all this applies to the octonionic projective plane, sadly, it does not apply to the bioctonionic plane! The bioctonionic plane does not come from a Euclidean Jordan algebra or a symmetric cone. Thus, to understand it as a space of pure states, we’d have to resort to a more general formalism.

There are a few papers that attempt exactly this:

  • Lawrence C. Biedenharn and Piero Truini, An 6𝒰(1)\mathcal{E}_6 \otimes \mathcal{U}(1) invariant quantum mechanics for a Jordan pair, Journal of Mathematical Physics 23 (1982), 1327-1345.

  • Lawrence C. Biedenharn and Piero Truini, Exceptional groups and elementary particle structures, Physica A: Statistical Mechanics and its Applications 14 (1982), 257–270.

  • Lawrence C. Biedenharn, G. Olivieri and Piero Truini, Three graded exceptional algebras and symmetric spaces, Zeitschrift Physik C — Particles and Fields 33 (1986), 47–65.

Here’s the basic idea. We can define 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}) to consist of 3×33 \times 3 hermitian matrices with entries in 𝕆\mathbb{C} \otimes \mathbb{O}, where ‘hermitian’ is defined using the star-algebra structure on 𝕆\mathbb{C} \otimes \mathbb{O} where we conjugate the octonion part but not the complex part! Then 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}) is just the complexification of 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}):

𝔥 3(𝕆)𝔥 3(𝕆) \mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}) \cong \mathbb{C} \otimes \mathfrak{h}_3(\mathbb{O})

Then because 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}) is a Jordan algebra over \mathbb{R}, 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}) is a Jordan algebra over \mathbb{C}. So we can do a lot with it. But it’s not a Euclidean Jordan algebra.

Puzzle. Show that it’s not.

So, Biedenharn and Truini need a different approach to relate 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}) to some sort of exotic quantum system. And they use an approach already known to mathematicians: namely, the theory of Jordan pairs! Here you work, not with a single element of 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{C} \otimes \mathbb{O}), but with a pair.

Jordan triple systems and Jordan pairs are two closely related generalizations of Jordan algebras. I’ve been working on the nLab articles about these concepts, so click the links if you want to learn more about them. I explain how either of these things gives you a 3-graded Lie algebra — that is, a \mathbb{Z}-graded Lie algebra that is nonvanishing only in the middle 3 grades:

𝔤=𝔤 1𝔤 0𝔤 1 \mathfrak{g} = \mathfrak{g}_{-1} \oplus \mathfrak{g}_0 \oplus \mathfrak{g}_1

And from a 3-graded Lie algebra you can get a symmetric space G/HG/H where the Lie algebra of GG is 𝔤\mathfrak{g} and the Lie algebra of HH is 𝔤 0\mathfrak{g}_0. Each tangent space of this symmetric space is thus isomorphic to 𝔤 1𝔤 1\mathfrak{g}_{-1} \oplus \mathfrak{g}_1 .

In the case relevant to the bioctonionic plane, the 3-graded Lie algebra is

𝔢 6=(𝕆)(𝔰𝔬(10))(𝕆) \mathfrak{e}_6 = (\mathbb{C} \otimes \mathbb{O}) \; \oplus \; (\mathfrak{so}(10) \oplus \mathbb{R}) \; \oplus \; (\mathbb{C} \otimes \mathbb{O})

So, the bioctonionic plane is a symmetric space on which E 6\text{E}_6 acts, with stabilizer group Spin(10)×U(1)\text{Spin}(10) \times \text{U}(1) (up to covering spaces), and with tangent space isomorphic to (𝕆) 2(\mathbb{C} \otimes \mathbb{O})^2.

So all this is potentially very nice. For much more on this theory, try the work of Ottmar Loos:

That’s a lot of stuff! For a quick overview of Loos’ work, I find this helpful:

Unfortunately Loos does not delve into examples, particularly the bioctonionic plane. For that, try Biedenharn and Truini, and also these:

To wrap things up, I should say a bit about ‘Hjelmslev planes’, since the bioctonionic plane is supposed to be one of these. Axiomatically, a Hjelmslev plane is a set PP of points, a set LL of lines, and an incidence relation between points and lines. We require that for any two distinct points there is at least one line incident to both, and for any two distinct line there is at least one point incident to both. If two points are incident to more than one line we say they are neighbors. If two lines are incident to more than one point we say they are neighbors. We demand that both these ‘neighbor’ relations are equivalence relations, and that if we quotient PP and LL by these equivalence relations, we get an axiomatic projective plane.

Challenge. What projective plane do we get if we apply this quotient construction to the bioctonionic plane?

My only guess is that we get the octonionic projective plane — but I don’t know why.

The literature on Hjelmslev planes seems a bit difficult, but I’m finding this to be a good introduction:

The answer to my puzzle should be here, because they’re talking about Hjelmslev planes built using split octonion algebras (like 𝕆\mathbb{C} \otimes \mathbb{O}):

  • Tonny A. Springer and Ferdinand D. Veldkamp, On Hjelmslev–Moufang planes, Mathematicsche Zeitschrift 107 (1968), 249–263.

But I don’t see the answer here yet!


  • Part 1. How to define octonion multiplication using complex scalars and vectors, much as quaternion multiplication can be defined using real scalars and vectors. This description requires singling out a specific unit imaginary octonion, and it shows that octonion multiplication is invariant under SU(3)\mathrm{SU}(3).
  • Part 2. A more polished way to think about octonion multiplication in terms of complex scalars and vectors, and a similar-looking way to describe it using the cross product in 7 dimensions.
  • Part 3. How a lepton and a quark fit together into an octonion — at least if we only consider them as representations of SU(3)\mathrm{SU}(3), the gauge group of the strong force. Proof that the symmetries of the octonions fixing an imaginary octonion form precisely the group SU(3)\mathrm{SU}(3).
  • Part 4. Introducing the exceptional Jordan algebra 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}): the 3×33 \times 3 self-adjoint octonionic matrices. A result of Dubois-Violette and Todorov: the symmetries of the exceptional Jordan algebra preserving their splitting into complex scalar and vector parts and preserving a copy of the 2×22 \times 2 adjoint octonionic matrices form precisely the Standard Model gauge group.
  • Part 5. How to think of 2×22 \times 2 self-adjoint octonionic matrices as vectors in 10d Minkowski spacetime, and pairs of octonions as left- or right-handed spinors.
  • Part 6. The linear transformations of the exceptional Jordan algebra that preserve the determinant form the exceptional Lie group E 6\mathrm{E}_6. How to compute this determinant in terms of 10-dimensional spacetime geometry: that is, scalars, vectors and left-handed spinors in 10d Minkowski spacetime.
  • Part 7. How to describe the Lie group E 6\mathrm{E}_6 using 10-dimensional spacetime geometry. This group is built from the double cover of the Lorentz group, left-handed and right-handed spinors, and scalars in 10d Minkowski spacetime.
  • Part 8. A geometrical way to see how E 6\mathrm{E}_6 is connected to 10d spacetime, based on the octonionic projective plane.
  • Part 9. Duality in projective plane geometry, and how it lets us break the Lie group E 6\mathrm{E}_6 into the Lorentz group, left-handed and right-handed spinors, and scalars in 10d Minkowski spacetime.
  • Part 10. Jordan algebras, their symmetry groups, their invariant structures — and how they connect quantum mechanics, special relativity and projective geometry.
  • Part 11. Particle physics on the spacetime given by the exceptional Jordan algebra: a summary of work with Greg Egan and John Huerta.
  • Part 12. The bioctonionic projective plane and its connections to algebra, geometry and physics.

Jordan EllenbergAnother example: sets of arcs and loops with bounded intersection

Yet another example of the kind of problem that falls under the sphere packing / cap set / Turan umbrella I blogged about last week.

Let G be SL_2(Z) and H be upper triangular matrices. Then G/H is identified with the set of pairs (a,b) in Z^2 with a,b relatively prime and a positive. Let m = 2. The orbits of G on (G/H)^2 are indexed by natural numbers: a pair (a,b),(c,d) of points in G/H is sent to the determinant |ad-bc|. If we take R to be the set {0,..,k}, then an R-set is a subset of Z^2 such that every one of these determinants has absolute value at most k. This is a very natural problem and it has a large literature: see e.g. this paper and this more recent one by Aougab and Gaster, which shows that an R-set has size at most k + O(sqrt(k) log k). But we know of no examples of an R-set of size larger than k+6!

Update: Blogging gets results! Two readers have informed me that this problem has recently been solved, and indeed, there is no R-set of size larger than k+6, and for large enough k, no R-set of size larger than k+4. And why do I specify that two readers informed me? Because there are two completely separate proofs! One by Kriepke and Schymura, and one by Balla, Filakovský, Kielak, Kráľ, and Schlomberg. It was actually a talk by Kráľ I saw this summer where I first learned about this problem; somehow I didn’t catch while preparing the post that he’d proved this theorem!

The two papers do not refer to each other. And in fact it looks like there are two completely separate groups of people working on this in parallel, and have been for years! There are the integral programming people, who think of the problem as the case r=2 of the question of r x n integer matrices with all rxr determinants bounded. And there are the topologists, who think of the problem as the case g=1 of the question of loops on a genus g Riemann surface with bounded pairwise intersection. The torus, where all fields of mathematics converge! The Wood Between The Worlds of math.

End of update, back to the original post.

This story generalizes. A coprime pair (a,b) in Z^2 is a loop on the torus, and |ad-bc| is the number of intersections between the loops (a,b) and (c,d). So we can ask: how many simple loops can there be on a genus g surface with no two having more than k intersections? This is an old question of Farb and Leininger when k=1 and the answer, proved by Aougab and Gaster this summer, is that the size grows at most quadratically in g. You can set this up as an R-set question for G the mapping class group of genus g. (Or you almost can; I think this question would ask about sets of simple loops which were all in the same orbit of the mapping class group.) For general k, the maximal size is expected to grow like g^{k+1}, and this is almost known. The case of a punctured disc is a theorem of Przyticki. The polynomial growth in g seems to me in the same spirit as what Guan and Ramos conjecture. The chromatic number version of this problem has also been studied.

Sorry if this is telegraphic! I just wanted to remind myself about the connection of this interesting body of problems in topology to the setup I mentioned earlier. I really do think that any interesting group (and the mapping class groups sure are interesting) produces interesting combinatorial questions via this machine.

We could also generalize the first question in a more number-theoretic, less low-dimensional topology direction, and ask: for a fixed m, how large can n be if there is an mxn matrix with primitive columns whose mxm minors all have determinant at most k?

December 03, 2025

n-Category Café log|x| + C revisited

A while ago on this blog, Tom posted a question about teaching calculus: what do you tell students the value of 1xdx\displaystyle\int \frac{1}{x}\,dx is? The standard answer is ln|x|+C\ln{|x|}+C, with CC an “arbitrary constant”. But that’s wrong if \displaystyle\int means (as we also usually tell students it does) the “most general antiderivative”, since

F(x)={ln|x|+C ifx<0 ln|x|+C + ifx>0 F(x) = \begin{cases} \ln{|x|} + C^- &\text{if}\;x\lt 0\\ \ln{|x|} + C^+ &\text{if}\;x\gt 0 \end{cases}

is a more general antiderivative, for two arbitrary constants C C^- and C +C^+. (I’m writing ln\ln for the natural logarithm function that Tom wrote as log\log, for reasons that will become clear later.)

In the ensuing discussion it was mentioned that other standard indefinite integrals like 1x 2dx=1x+C\displaystyle\int \frac{1}{x^2}\,dx = -\frac{1}{x} + C are just as wrong. This happens whenever the domain of the integrand is disconnected: the “arbitrary constant” CC is really only locally constant. Moreover, Mark Meckes pointed out that believing in such formulas can lead to mistaken calculations such as

1 11x 2dx=1x] 1 1=2 \int_{-1}^1 \frac{1}{x^2}\,dx = \left.-\frac{1}{x}\right]_{-1}^1 = -2

which is “clearly nonsense” since the integrand is everywhere positive.

In this post I want to argue that there’s actually a very natural perspective from which 1x 2dx=1x+C\displaystyle\int \frac{1}{x^2}\,dx = -\frac{1}{x} + C is correct, while 1xdx=ln|x|+C\displaystyle\int \frac{1}{x}\,dx = \ln{|x|}+C is wrong for a different reason.

The perspective in question is complex analysis. Most of the functions encountered in elementary calculus are actually complex-analytic — the only real counterexamples are explicit “piecewise” functions and things like |x|{|x|}, which are mainly introduced as counterexamples to illustrate the meaning of continuity and differentiability. Therefore, it’s not unreasonable to interpret the indefinite integral f(x)dx\displaystyle\int f(x)\,dx as asking for the most general complex-analytic antiderivative of ff. And the complex domain of 1z\frac{1}{z} and 1z 2\frac{1}{z^2} is {0}\mathbb{C}\setminus \{0\}, which is connected!

Thus, for instance, since ddz[1z]=1z 2\frac{d}{d z}\left[-\frac{1}{z}\right] = \frac1{z^2}, it really is true that the most general (complex-analytic) antiderivative of 1z 2\frac{1}{z^2} is 1z+C-\frac{1}{z}+C for a single arbitrary constant CC, so we can write 1z 2dz=1z+C\displaystyle\int \frac{1}{z^2}\,dz = -\frac{1}{z} + C. Note that any such antiderivative has the same domain {0}\mathbb{C}\setminus \{0\} as the original function.

In addition, the dodgy calculation

1 11z 2dz=1z] 1 1=2 \int_{-1}^1 \frac{1}{z^2}\,dz = \left.-\frac{1}{z}\right]_{-1}^1 = -2

is actually correct if we interpret 1 1\int_{-1}^1 to mean the integral along some (in fact, any) curve in \mathbb{C} from 1-1 to 11 that doesn’t pass through the singularity z=0z=0. Of course, this doesn’t offend against signs because any such path must pass through non-real numbers, whose squares can contribute negative real numbers to the integral.

The case of 1zdz\displaystyle\int \frac{1}{z}\,dz is a bit trickier, because the complex logarithm is multi-valued. However, if we’re willing to work with multi-valued functions (which precisely means functions whose domain is a Riemann surface covering some domain in \mathbb{C}), we have such a multi-valued function that I’ll denote log\log (in contrast to the usual real-number function ln\ln) defined on a connected domain, and there we have ddz[log(z)]=1z\frac{d}{d z}\left[ \log(z) \right] = \frac{1}{z}. Thus, the most general (complex-analytic) antiderivative of 1z\frac{1}{z} is log(z)+C\log(z)+C where CC is a single arbitrary constant, so we can write 1zdz=log(z)+C\displaystyle\int \frac{1}{z}\,dz = \log(z) + C.

What happened to ln|x|\ln{|x|}? Well, as it happens, if xx is a negative real number and LogLog denotes the principal branch of the complex logarithm, then Log(x)=ln|x|+iπLog(x) = \ln{|x|} + i\pi, hence ln|x|=Log(x)iπ\ln{|x|} = Log(x) - i\pi. Therefore, the antiderivative ln|x|\ln{|x|} for negative real xx is of the form Log(x)+C\Log(x)+C, where Log\Log is a branch of the complex logarithm and CC is a constant (namely, iπ-i\pi).

Of course it is also true that for positive real xx, the antiderivative ln|x|=ln(x)\ln{|x|} = \ln(x) is of the form Log(x)+C\Log(x)+C for some constant CC, but in this case the constant is 00. And changing the branch of the logarithm changes the constant by 2iπ2i\pi, so it can never make the constants 00 and iπ-i\pi coincide. Thus, unlike 1x+C-\frac{1}{x} +C, the real-number function ln|x|+C\ln{|x|} + C in the “usual answer” is not the restriction to {0}\mathbb{R}\setminus \{0\} of any complex-analytic antiderivative of 1x\frac{1}{x} on a connected domain. This is what I mean by saying that 1xdx=ln|x|+C\displaystyle\int \frac{1}{x}\,dx = \ln{|x|}+C is now wrong for a different reason. And we can see that the analogous dodgy calculation

1 11xdx=ln|x|] 1 1=0 \int_{-1}^1 \frac{1}{x}\,dx = \ln{|x|}\Big]_{-1}^1 = 0

is also still wrong. If γ\gamma is a path from 1-1 to 11 in {0}\mathbb{C}\setminus \{0\}, the value of γ1xdx\int_{\gamma} \frac{1}{x}\,dx depends on γ\gamma, but it never equals 00: it’s always an odd integer multiple of iπi\pi, depending on how many times γ\gamma winds around the origin.

I’m surprised that no one in the previous discussion, including me, brought this up. Of course we probably don’t want to teach our elementary calculus students complex analysis (although I’m experimenting with introducing some complex numbers in second-semester calculus). But this perspective makes me less unhappy about writing 1x 2dx=1x+C\displaystyle\int \frac{1}{x^2}\,dx = -\frac{1}{x} + C and 1xdx=log(x)+C\displaystyle\int \frac{1}{x}\,dx = \log(x) + C (no absolute value!).

Tommaso DorigoConferences Good And Bad, In A Profit-Driven Society

Nowadays researchers and scholars of all ages and specialization find themselves struggling with mailboxes pestered with invitations to conferences, invitations to submit papers to journals, invitations to participate in the editorial board of journals, invitations to receive prizes for this or that reason; and of course, 99% of the origin of these invitations are individuals running fake conferences, scam, or predatory journals. Spam filters are not extremely good at distinguishing good and bad invitations, so if one wants to avoid discarding prestigious opportunities the only option is a painful manual screening.

read more

Terence TaoClimbing the cosmic distance ladder: another sample chapter

Five years ago, I announced a popular science book project with Tanya Klowden on the cosmic distance ladder, in which we released a sample draft chapter of the book, covering the “fourth rung” of the ladder, which for us meant the distances to the planets. In the intervening time, a number of unexpected events have slowed down this project significantly; but I am happy to announce that we have completed a second draft chapter, this time on the “seventh rung” of measuring distances across the Milky Way, which required the maturation of the technologies of photography and spectroscopy, as well as the dawn of the era of “big data” in the early twentieth century, as exemplified for instance by the “Harvard computers“.

We welcome feedback of course, and are continuing to work to complete the book despite the various delays. In the mean time, you can check out our instagram account for the project, or the pair of videos that Grant Sanderson (3blue1brown) produced with us on this topic, which I previously blogged about here.

Thanks to Clio Cresswell, Riley Tao, and Noah Klowden for comments and corrections.

December 02, 2025

Terence TaoQuantitative correlations and some problems on prime factors of consecutive integers

Joni Teravainen and I have uploaded to the arXiv our paper “Quantitative correlations and some problems on prime factors of consecutive integers“. This paper applies modern analytic number theory tools – most notably, the Maynard sieve and the recent correlation estimates for bounded multiplicative functions of Pilatte – to resolve (either partially or fully) some old problems of Erdős, Strauss, Pomerance, Sárközy, and Hildebrand, mostly regarding the prime counting function

\displaystyle  \omega(n) := \sum_{p|n} 1

and its relatives. The famous Hardy–Ramanujan and Erdős–Kac laws tell us that asymptotically for {n \sim x}, {\omega(n)} should behave like a gaussian random variable with mean and variance both close to {\log\log x}; but the question of the joint distribution of consecutive values such as {\omega(n), \omega(n+1)} is still only partially understood. Aside from some lower order correlations at small primes (arising from such observations as the fact that precisely one of {n,n+1} will be divisible by {2}), the expectation is that such consecutive values behave like independent random variables. As an indication of the state of the art, it was recently shown by Charamaras and Richter that any bounded observables {f(\omega(n))}, {g(\omega(n+1))} will be asymptotically decorrelated in the limit {n \rightarrow \infty} if one performs a logarithmic statistical averaging. Roughly speaking, this confirms the independence heuristic at the scale {\sqrt{\log\log x}} of the standard deviation, but does not resolve finer-grained information, such as precisely estimating the probability of the event {\omega(n)=\omega(n+1)}.

Our first result, answering a question of Erdős, shows that there are infinitely many {n} for which one has the bound

\displaystyle  \omega(n+k) \ll k

for all {k \geq 1}. For {k \gg \log\log n}, such a bound is already to be expected (though not completely universal) from the Hardy–Ramanujan law; the main difficulty is thus with the short shifts {k = o(\log\log n)}. If one only had to demonstrate this type of bound for a bounded number of {k}, then this type of result is well within standard sieve theory methods, which can make any bounded number of shifts {n+k} “almost prime” in the sense that {\omega(n+k)} becomes bounded. Thus the problem is that the “sieve dimension” {\sim \log\log n} grows (slowly) with {n}. When writing about this problem in 1980, Erdős and Graham write “we just know too little about sieves to be able to handle such a question (“we” here means not just us but the collective wisdom (?) of our poor struggling human race)”.

However, with the advent of the Maynard sieve (also sometimes referred to as the Maynard–Tao sieve), it turns out to be possible to sieve for the conditions {\omega(n+k) \ll k} for all {k = o(\log\log n)} simultaneously (roughly speaking, by sieving out any {n} for which {n+k} is divisible by a prime {p \ll x^{1/\exp(Ck)}} for a large {C}), and then performing a moment calculation analogous to the standard proof (due to Turán) of the Hardy–Ramanujan law, but weighted by the Maynard sieve. (In order to get good enough convergence, one needs to control fourth moments as well as second moments, but these are standard, if somewhat tedious, calculations).

Our second result, which answers a separate question of Erdős, establishes that the quantity

\displaystyle  \sum_n \frac{\omega(n)}{2^n} = 0.5169428\dots

is irrational; this had recently been established by Pratt under the assumption of the prime tuples conjecture, but we are able to establish this result unconditoinally. The binary expansion of this number is of course closely related to the distribution of {\omega}, but in view of the Hardy–Ramanujan law, the {n^{th}} digit of this number is influenced by about {\log\log\log n} nearby values of {\omega}, which is too many correlations for current technology to handle. However, it is possible to do some “Gowers norm” type calculations to decouple things to the point where pairwise correlation information is sufficient. To see this, suppose for contradiction that this number was a rational {a/q}, thus

\displaystyle  q \sum_n \frac{\omega(n)}{2^n} = 0 \hbox{ mod } 1.

Multiplying by {2^n}, we obtain some relations between shifts {\omega(n+h)}:

\displaystyle  q \sum_{h=1}^\infty \frac{\omega(n+h)}{2^h} = 0 \hbox{ mod } 1.

Using the additive nature of {\omega}, one then also gets similar relations on arithmetic progressions, for many {n} and {p}:

\displaystyle  q \sum_{h=1}^\infty \frac{\omega(n+ph)}{2^h} = 0 \hbox{ mod } 1.

Taking alternating sums of this sort of identity for various {n} and {p} (in analogy to how averages involving arithmetic progressions can be related to Gowers norm-type expressions over cubes), one can eventually arrive eliminate the contribution of small {H}, and arrive at an identity of the form

\displaystyle  q \sum_{h=1}^\infty \frac{\sum_{\epsilon \in \{0,1\}^K} (-1)^{|\epsilon|} \omega(n + r_{\epsilon,h+K})}{2^{h+K}} = 0 \hbox{ mod } 1 \ \ \ \ \ (1)

for many {n}, where {K} is a parameter (we eventually take {K \sim \log\log\log\log n}) and {r_{\epsilon,h+K}} are various shifts that we will not write out explicitly here. This looks like quite a messy expression; however, one can adapt proofs of the Erdős–Kac law and show that, as long as one ignores the contribution of really large prime factors (of order {\gg n^{1/10}}, say) to the {\omega(n + r_{\epsilon,h+K})}, that this sort of sum behaves like a gaussian, and in particular once one can show a suitable local limit theorem, one can contradict (1). The contribution of the large prime factors does cause a problem though, as a naive application of the triangle inequality bounds this contribution by {O(1)}, which is an error that overwhelms the information provided by (1). To resolve this we have to adapt the pairwise correlation estimates of Pilatte mentioned earlier to demonstrate that the these contributions are in fact {o(1)}. Here it is important that the error estimates of Pilatte are quite strong (of order {O(\log^{-c} n)}); previous correlation estimates of this type (such as those used in this earlier paper with Joni) turn out to be too weak for this argument to close.

Our final result concerns the asymptotic behavior of the density

\displaystyle  \frac{1}{x} \{n \leq x: \omega(n+1) = \omega(n)\}

(we also address similar questions for {\Omega(n+1)=\Omega(n)} and {\tau(n+1)=\tau(n)}). Heuristic arguments led Erdős, Pomerance, and Sárközy to conjecture that this quantity was asymptotically {\frac{1}{2\sqrt{\pi \log\log x}}}. They were able to establish an upper bound of {O(1/\log\log x)}, while Hildebrand obtained a lower bound of {\gg 1/(\log\log x)^3}, due to Hildebrand. Here, we obtain the asymptotic for almost all {x} (the limitation here is the standard one, which is that the current technology on pairwise correlation estimates either requires logarithmic averaging, or is restricted to almost all scales rather than all scales). Roughly speaking, the idea is to use the circle method to rewrite the above density in terms of expressions

\displaystyle  \frac{1}{x} \sum_{n \leq x} e(\alpha \omega(n+1)) e(-\alpha \omega(n))

for various frequencies {\alpha}, use the estimates of Pilatte to handle the minor arc {\alpha}, and convert the major arc contribution back into physical space (in which {\omega(n+1)} and {\omega(n)} are now permitted to differ by a large amount) and use more traditional sieve theoretic methods to estimate the result.

Jordan Ellenberg“That’s an excellent question!”

Most commercial large language models are tuned to butter you up a little, or a lot, whenever you submit a query. “That’s an excellent question!” Well, OK, I guess I thought it was all right or I wouldn’t have typed it in the window. But I cannot deny there’s something in me that authentically enjoys the predetermined praise. The feeling reminded me of an artwork I saw a few years ago, which I blogged about, but that post is short so maybe I’ll just reproduce it here:


I gave a talk at Williams College last year and took a little while to visit one of my favorite museums, Mass MoCA. There’s a new installation there, by Taryn Simon, called Assembled Audience. You walk in through a curtained opening and you’re in a pitch-black space. It’s very quiet. And then, slowly, applause starts to build. Bigger and bigger. About a minute of swell until the invisible crowd out there in the dark is going absolutely fucking nuts.

And I have to be honest, whatever this may say about me: I felt an incredible warmth and safety and satisfaction, standing there, being clapped for and adored by a recording of a crowd. Reader, I stayed for a second cycle.


Is this feeling universal, or are we just talking about my particular character flaw here?

The Instagram page for the most recent showing of Assembled Audience, in Melbourne last fall, has fascinating comments:

“Terrifying, I loved it.”

“I felt like they were clapping for me, I could’ve stayed in there forever!”

“I loved the dark space, it was so scary!”

“I loved the clapping space, I wish I could have that as my wake up alarm.”

“The most excellent and interesting exhibition pieces, I felt alive in the dark room.”

December 01, 2025

John BaezSummer Research at Topos

You can now apply for the 2026 Summer Research Associate program at the Topos Institute! This is a great opportunity.

Details and instructions on how to apply are in the official announcement.

A few important points:

• The application deadline is January 16, 2026.
• The position is paid and in-person in Berkeley, California.

These positions will last for 8 – 10 weeks, starting in June 2026 and ending in August. Each position will be mentored by Topos research staff or a select number of invited mentors. All positions are 40 hours/week, and the salary starts at $30-$50/hour.

There’s a research track and an engineering track. For the research track, possible topics include:

• Computational category theory using CatColab (Rust/Typescript skills recommended)
• Double category theory
• Categorical statistics
• Polynomial functors
• Interacting dynamical systems
• Hybrid dynamical systems, attractor theory and fast-slow dynamics
• Proof assistants, formal verification, or structure editors
• Philosophical and ethical aspects of applied category theory

For the engineering track, possible topics include:

• Delivery and support of mathematical technologies for various scientific disciplines and applications, and/or analysis, documentation, or guidance on their uses.
• Designing, implementing, testing, and maintaining software at the Topos Institute, in close collaboration with the research staff and in line with institute’s scientific strategy and mission.
• Contributing to developing the CatColab platform, including front end development in TypeScript and/or back end development in Rust. You might also contribute to the mathematical core, written in Rust, as your mathematical experience permits.

All positions require collaboration within a multi-disciplinary research environment. Each summer research associate will complete a specific Topos project, and will write a blog post by the last week of their employment. These projects may include an internal talk, software contribution, or paper. Go here to see the accomplishments of previous research associates.

Topos is committed to building a team with diverse perspectives and life experiences, so those with personal or professional backgrounds underrepresented at Topos are highly encouraged to apply. They are dedicated to shaping the future of technology to ensure a more equitable and just world, and believe that a technology that supports a healthy society can only be built by an organization that supports its team members.

Secret Blogging SeminarCongress proposes cutting of all funding to US academics who mentor Chinese students

I’m writing to point out a potential law which should be gathering more opposition and attention in math academia: The Securing American Funding and Expertise from Adversarial Research Exploitation Act. This is an amendment to the 2026 National Defense Authorization Act which has passed the House and could be added to the final version of the bill during reconcilliation in the Senate. I’m pulling most of my information from an article in Science.

This act would ban any US scientist from receiving federal funding if they have, within the last five years, worked with anyone from China, Russia, Iran or North Korea, where “worked with” includes joint research, co-authorship on papers, or advising a foreign graduate student or postdoctoral fellow. As I said in my message to my senators, this is everyone. Every mathematician has advised Chinese graduate students or collaborated with Chinese mathematicians, because China is integrated into the academic world and is one fifth of the earth.

This obviously isn’t secret, since you can read about it in Science, but I am surprised that I haven’t heard more alarm. Obvious people to contact are your senators and your representatives. I would also suggest contacting members of the Senate armed services committee, who are in charge of reconciling the House and Senate versions of the bill.

November 30, 2025

Scott Aaronson Mihai Pătrașcu Best Paper Award: Guest post from Seth Pettie

Scott’s foreword: Today I’m honored to turn over Shtetl-Optimized to a guest post from Michigan theoretical computer scientist Seth Pettie, who writes about a SOSA Best Paper Award newly renamed in honor of the late Mihai Pătrașcu. Mihai, who I knew from his student days, was a brash, larger-than-life figure in theoretical computer science, for a brief few years until brain cancer tragically claimed him at the age of 29. Mihai and I didn’t always agree—indeed, I don’t think he especially liked me, or this blog—but as I wrote when he passed, his death made any squabbles seem trivial in retrospect. He was a lion of data structures, and it’s altogether fitting that this award be named for him. –SA


Seth’s guest post:

The SIAM Symposium on Simplicity in Algorithms (SOSA) was created in 2018 and has been awarding a Best Paper Award since 2020. This year the Steering Committee renamed this award after Mihai Pătrașcu, an extraordinary researcher in theoretical computer science who passed away before his time, in 2012.

Mihai’s research career lasted just a short while, from 2004-2012, but in that span of time he had a huge influence on research in geometry, graph algorithms, data structures, and especially lower bounds. He revitalized the entire areas of cell-probe lower bounds and succinct data structures, and laid the foundation for fine-grained complexity with the first 3SUM-hardness proof for graph problems. He lodged the most successful attack to date on the notorious dynamic optimality conjecture, then recast it
as a pure geometry problem. If you are too young to have met Mihai personally, I encourage you to pick up one of his now-classic papers. They are a real joy to read—playful and full of love for theoretical computer science.

The premise of SOSA is that simplicity is extremely valuable, rare, and inexplicably undervalued. We wanted to create a venue where the chief metrics of success were simplicity and insight. It is fitting that the SOSA Best Paper Award be named after Mihai. He brought “fresh eyes” to every problem he worked on, and showed that the cure for our problems is usually one key insight (and of course some mathematical gymnastics).

Let me end by thanking the SOSA 2026 Program Committee, co-chaired by Sepehr Assadi and Eva Rotenberg, and congratulating the authors of the SOSA 2026 Mihai Pătrașcu Best Paper:

This award will be given at the SODA/SOSA business meeting in Vancouver, Canada, on January 12, 2026.

n-Category Café Beyond the Geometry of Music

Yesterday I had a great conversation with Dmitri Tymoczko about groupoids in music theory. But at this Higgs Centre Colloquium, he preferred to downplay groupoids and talk in a way physicists would enjoy more. Click here to watch his talk!

What’s great is that Tymoczkyo not faking it: he’s really found deep ways in which symmetry shows up pervasively in music.

At first he tried to describe them geometrically using orbifolds, which are spaces in which some singular points have nontrivial symmetry groups, like the tip of a cone formed by modding out the plane by the action of the group /n\mathbb{Z}/n. But then he realized that the geometry was less important than the symmetry, which you can describe using groupoids. That’s why his talk is called “Beyond the geometry of music”.

I’m helping him with his work on groupoids, and I hope he explains his work to mathematicians someday without pulling his punches. I didn’t get to interview him yesterday, but I’ll try to do that soon.

For now you can read his books A Geometry of Music and Harmony: an Owner’s Manual along with many papers. What I’ve read so far is really exciting.

November 29, 2025

Doug NatelsonWhat is the orbital Hall effect?

In the course of thinking about how best to revise my too-math-infused post about quantum geometry, I realized that writing about the orbital Hall effect lays nice groundwork.  

I've previously written about the spin Hall effect (SHE), in which a charge current \(\mathbf{j}_{\mathrm{c}}\) directed along \(\hat{\mathbf{x}}\) generates a net flow of \(\hat{\mathbf{y}}\)-directed spin angular momentum along the \(\hat{\mathbf{z}}\) direction.  This is a consequence of spin-orbit coupling, and it was first predicted in 1971 with a major revival sparked in 1999.  Electrically generating angular momentum currents has proven very useful, leading to many ideas about magnetic memory devices.  Microscopically, it's not easy to develop an intuition about the SHE, though as a spin-orbit effect, it is expected to be much stronger in heavier metals, since the spin-orbit coupling in atomic orbitals scales like \(Z^{4}\), and electronic bands in solids are built from those orbitals.  

That fact, that the electronic bands originate from atomic orbitals, is something that can get lost in a Bloch wave/nearly-free electron treatment of electronic structure.  In the orbital Hall effect, this idea is paramount.  This was explained clearly in this PRL (arXiv here).  The little \(p\)-orbitals are drawn on top of the \(k_{x}-k_{y}\) plane, to illustrate the idea that the electronic states in \(\mathbf{k}\)-space have different orbital content, depending on \(\mathbf{k}\).   The blue circle represents the "Fermi disk", with \(\mathbf{k}\)-states inside the circle occupied, and \(\mathbf{k}\)-states outside the circle empty.  
Adapted from Fig. 1 here.

When no electric field is applied, the Fermi disk is centered on \(\mathbf{k} = 0\); there is no net current, and there is no net orbital angular momentum once all the filled states are considered.  When an electric field is applied in the \(+x\) direction, though, the Fermi disk is shifted away from the origin in the \(-x\) direction (because of our convention that electrons are negatively charged).  Now adding up the \(z\)-directed orbital angular momentum contained within the Fermi disk, there is net \(+z\) orbital angular momentum carried by states with positive \(k_{y}\), and net \(-z\) orbital angular momentum carried by states with negative \(k_{y}\).  So, for this orbital texture, a charge current \(\mathbf{j}_{\mathrm{c}}\) directed along \(+\hat{\mathbf{x}}\) generates a net flow of \(\hat{\mathbf{z}}\)-directed orbital angular momentum along the \(+\hat{\mathbf{y}}\) direction.  Charge current generates a transverse flow of orbital angular momentum, entirely due to the way atomic orbitals come together to make Bloch states in \(\mathbf{k}\)-space, independent of any spin-orbit physics.  That's why the orbital Hall effect has been inferred experimentally in several materials with weak spin-orbit effects, like chromium and titanium.

These effects can be large, and orbital Hall physics plus some \(\mathbf{L}\cdot\mathbf{S}\) coupling may be responsible for some of the results labeled as spin Hall.  See here for a discussion.  Electrically pumping around angular momentum through orbital and spin Hall effects, and their inverses, is the idea behind a variety of device concepts for memory (e.g. here) and logic.  Fun stuff.








November 28, 2025

Matt von HippelBonus Info For “Cosmic Paradox Reveals the Awful Consequence of an Observer-Free Universe”

I had a piece in Quanta Magazine recently, about a tricky paradox that’s puzzling quantum gravity researchers and some early hints at its resolution.

The paradox comes from trying to describe “closed universes”, which are universes where it is impossible to reach the edge, even if you had infinite time to do it. This could be because the universe wraps around like a globe, or because the universe is expanding so fast no traveler could ever reach an edge. Recently, theoretical physicists have been trying to describe these closed universes, and have noticed a weird issue: each such universe appears to have only one possible quantum state. In general, quantum systems have more possible states the more complex they are, so for a whole universe to have only one possible state is a very strange thing, implying a bizarrely simple universe. Most worryingly, our universe may well be closed. Does that mean that secretly, the real world has only one possible state?

There is a possible solution that a few groups are playing around with. The argument that a closed universe has only one state depends on the fact that nothing inside a closed universe can reach the edge. But if nothing can reach the edge, then trying to observe the universe as a whole from outside would tell you nothing of use. Instead, any reasonable measurement would have to come from inside the universe. Such a measurement introduces a new kind of “edge of the universe”, this time not in the far distance, but close by: the edge between an observer and the rest of the world. And when you add that edge to the calculations, the universe stops being closed, and has all the many states it ought to.

This was an unusually tricky story for me to understand. I narrowly avoided several misconceptions, and I’m still not sure I managed to dodge all of them. Likewise, it was unusually tricky for the editors to understand, and I suspect it was especially tricky for Quanta’s social media team to understand.

It was also, quite clearly, tricky for the readers to understand. So I thought I would use this post to clear up a few misconceptions. I’ll say a bit more about what I learned investigating this piece, and try to clarify what the result does and does not mean.

Q: I’m confused about the math terms you’re using. Doesn’t a closed set contain its boundary?

A: Annoyingly, what physicists mean by a closed universe is a bit different from what mathematicians mean by a closed manifold, which is in turn more restrictive than what mathematicians mean by a closed set. One way to think about this that helped me is that in an open set you can take a limit that takes you out of the set, which is like being able to describe a (possibly infinite) path that takes you “out of the universe”. A closed set doesn’t have that, every path, no matter how long, still ends up in the same universe.

Q: So a bunch of string theorists did a calculation and got a result that doesn’t make sense, a one-state universe. What if they’re just wrong?

A: Two things:

First, the people I talked to emphasized that it’s pretty hard to wiggle out of the conclusion. It’s not just a matter of saying you don’t believe in string theory and that’s that. The argument is based in pretty fundamental principles, and it’s not easy to propose a way out that doesn’t mess up something even more important.

That’s not to say it’s impossible. One of the people I interviewed, Henry Maxfield, thinks that some of the recent arguments are misunderstanding how to use one of their core techniques, in a way that accidentally presupposes the one-state universe.

But even he thinks that the bigger point, that closed universes have only one state, is probably true.

And that’s largely due to a second reason: there are older arguments that back the conclusion up.

One of the oldest dates back to John Wheeler, a physicist famous for both deep musings about the nature of space and time and coining evocative terms like “wormhole”. In the 1960’s, Wheeler argued that, in a theory where space and time can be curved, one should think of a system’s state as including every configuration it can evolve into over time, since it can be tricky to specify a moment “right now”. In a closed universe, you could expect a quantum system to explore every possible configuration…meaning that such a universe should be described by only one state.

Later, physicists studying holography ran into a similar conclusion. They kept noticing systems in quantum gravity where you can describe everything that happens inside by what happens on the edges. If there are no edges, that seems to suggest that in some sense there is nothing inside. Apparently, Lenny Susskind had a slide at the end of talks in the 90’s where he kept bringing up this point.

So even if the modern arguments are wrong, and even if string theory is wrong…it still looks like the overall conclusion is right.

Q: If a closed universe has only one state, does that make it deterministic, and thus classical?

A: Oh boy…

So, on the one hand, there is an idea, which I think also goes back to Wheeler, that asks: “if the universe as a whole has a wavefunction, how does it collapse?” One possibility is that the universe has only one state, so that nobody is needed to collapse the wavefunction, it already is in a definite state.

On the other hand, a universe with only one state does not actually look much like a classical universe. Our universe looks classical largely due to a process called decoherence, where small quantum systems interact with big quantum systems with many states, diluting quantum effects until the world looks classical. If there is only one state, there are no big systems to interact with, and the world has large quantum fluctuations that make it look very different from a classical universe.

Q: How, exactly, are you defining “observer”?

A: A few commenters helpfully chimed in to talk about how physics models observers as “witness” systems, objects that preserve some record of what happens to them. A simple example is a ball sitting next to a bowl: if you find the ball in the bowl later, it means something moved it. This process, preserving what happens and making it more obvious, is in essence how physicists think about observers.

However, this isn’t the whole story in this case. Here, different research groups introducing observers are doing it in different ways. That’s, in part, why none of them are confident they have the right answer.

One of the approaches describes an observer in terms of its path through space and time, its worldline. Instead of a detailed witness system with specific properties, all they do is pick out a line and say “the observer is there”. Identifying that line, and declaring it different from its surroundings, seems to be enough to recover the complexity the universe ought to have.

The other approach treats the witness system in a bit more detail. We usually treat an observer in quantum mechanics as infinitely large compared to the quantum systems they measure. This approach instead gives the observer a finite size, and uses that to estimate how far their experience will be from classical physics.

Crucially, both approaches aren’t a matter of defining a physical object, and looking for it in the theory. Given a collection of atoms, neither team can tell you what is an observer, and what isn’t. Instead, in each approach, the observer is arbitrary: a choice, made by us when we use quantum mechanics, of what to count as an observer and what to count as the rest of the world. That choice can be made in many different ways, and each approach tries to describe what happens when you change that choice.

This is part of what makes this approach uncomfortable to some more philosophically-minded physicists: it treats observers not as a predictable part of the physical world, but as a mathematical description used to make statements about the world.

Q: If these ideas come from AdS/CFT, which is an open universe, how do you use them to describe a closed universe?

A: While more examples emerged later, initially theorists were thinking about two types of closed universes:

First, think about a black hole. You may have heard that when you fall into a black hole, you watch the whole universe age away before your eyes, due to the dramatic differences in the passage of time caused by the extreme gravity. Once you’ve seen the outside universe fade away, you are essentially in a closed universe of your own. The outside world will never affect you again, and you are isolated, with no path to the outside. These black hole interiors are one of the examples theorists looked at.

The other example are so-called “baby universes”. When physicists use quantum mechanics to calculate the chance of something happening, they have to add up every possible series of events that could have happened in between. For quantum gravity, this includes every possible arrangement of space and time. This includes arrangements with different shapes, including ones with tiny extra “baby universes” which branch off from the main universe and return. Universes with these “baby universes” are another example that theorists considered to understand closed universes.

Q: So wait, are you actually saying the universe needs to be observed to exist? That’s ridiculous, didn’t the universe exist long before humans existed to observe it? Is this some sort of Copenhagen Interpretation thing, or that thing called QBism?

You’re starting to ask philosophical questions, and here’s the thing:

There are physicists who spend their time thinking about how to interpret quantum mechanics. They talk to philosophers, and try to figure out how to answer these kinds of questions in a consistent and systematic way, keeping track of all the potential pitfalls and implications. They’re part of a subfield called “quantum foundations”.

The physicists whose work I was talking about in that piece are not those people.

Of the people I interviewed, one of them, Rob Myers, probably has lunch with quantum foundations researchers on occasion. The others, based at places like MIT and the IAS, probably don’t even do that.

Instead, these are people trying to solve a technical problem, people whose first inclination is to put philosophy to the side, and “shut up and calculate”. These people did a calculation that ought to have worked, checking how many quantum states they could find in a closed universe, and found a weird and annoying answer: just one. Trying to solve the problem, they’ve done technical calculation work, introducing a path through the universe, or a boundary around an observer, and seeing what happens. While some of them may have their own philosophical leanings, they’re not writing works of philosophy. Their papers don’t talk through the philosophical implications of their ideas in all that much detail, and they may well have different thoughts as to what those implications are.

So while I suspect I know the answers they would give to some of these questions, I’m not sure.

Instead, how about I tell you what I think?

I’m not a philosopher, I can’t promise my views will be consistent, that they won’t suffer from some pitfall. But unlike other people’s views, I can tell you what my own views are.

To start off: yes, the universe existed before humans. No, there is nothing special about our minds, we don’t have psychic powers to create the universe with our thoughts or anything dumb like that.

What I think is that, if we want to describe the world, we ought to take lessons from science.

Science works. It works for many reasons, but two important ones stand out.

Science works because it leads to technology, and it leads to technology because it guides actions. It lets us ask, if I do this, what will happen? What will I experience?

And science works because it lets people reach agreement. It lets people reach agreement because it lets us ask, if I observe this, what do I expect you to observe? And if we agree, we can agree on the science.

Ultimately, if we want to describe the world with the virtues of science, our descriptions need to obey this rule: they need to let us ask “what if?” questions about observations.

That means that science cannot avoid an observer. It can often hide the observer, place them far away and give them an infinite mind to behold what they see, so that one observer is essentially the same as another. But we shouldn’t expect to always be able to do this. Sometimes, we can’t avoid saying something about the observer: about where they are, or how big they are, for example.

These observers, though, don’t have to actually exist. We should be able to ask “what if” questions about others, and that means we should be able to dream up fictional observers, and ask, if they existed, what would they see? We can imagine observers swimming in the quark-gluon plasma after the Big Bang, or sitting inside a black hole’s event horizon, or outside our visible universe. The existence of the observer isn’t a physical requirement, but a methodological one: a restriction on how we can make useful, scientific statements about the world. Our theory doesn’t have to explain where observers “come from”, and can’t and shouldn’t do that. The observers aren’t part of the physical world being described, they’re a precondition for us to describe that world.

Is this the Copenhagen Interpretation? I’m not a historian, but I don’t think so. The impression I get is that there was no real Copenhagen Interpretation, that Bohr and Heisenberg, while more deeply interested in philosophy than many physicists today, didn’t actually think things through in enough depth to have a perspective you can name and argue with.

Is this QBism? I don’t think so. It aligns with some things QBists say, but they say a lot of silly things as well. It’s probably some kind of instrumentalism, for what that’s worth.

Is it logical positivism? I’ve been told logical positivists would argue that the world outside the visible universe does not exist. If that’s true, I’m not a logical positivist.

Is it pragmatism? Maybe? What I’ve seen of pragmatism definitely appeals to me, but I’ve seen my share of negative characterizations as well.

In the end, it’s an idea about what’s useful and what’s not, about what moves science forward and what doesn’t. It tries to avoid being preoccupied with unanswerable questions, and as much as possible to cash things out in testable statements. If I do this, what happens? What if I did that instead?

The results I covered for Quanta, to me, show that the observer matters on a deep level. That isn’t a physical statement, it isn’t a mystical statement. It’s a methodological statement: if we want to be scientists, we can’t give up on the observer.

Tommaso DorigoUSERN: 10 Years Of Non-Profit Action Supporting Science Education And Research

The 10th congress of the USERN organization was held on November 8-10 in Campinas, Brazil. Some time has gone by, so it is due time for me to report on the event. I could not attend in person for a cause of force majeure, but I was connected via zoom, and I also delivered two recorded speeches plus one talk in one of the parallel "virtual session" that were run via zoom in the evenings (CET) after the in-person program of the day was over. 

read more

November 27, 2025

John PreskillWhat distinguishes quantum from classical thermodynamics?

Should you require a model for an Oxford don in a play or novel, look no farther than Andrew Briggs. The emeritus professor of nanomaterials speaks with a southern-English accent as crisp as shortbread, exhibits manners to which etiquette influencer William Hanson could aspire, and can discourse about anything from Bantu to biblical Hebrew. I joined Andrew for lunch at St. Anne’s College, Oxford, this month.1 Over vegetable frittata, he asked me what unifying principle distinguishes quantum from classical thermodynamics.

With a thermodynamic colleague at the Oxford University Museum of Natural History

I’d approached quantum thermodynamics from nearly every angle I could think of. I’d marched through the thickets of derivations and plots; I’d journeyed from subfield to subfield; I’d gazed down upon the discipline as upon a landscape from a hot-air balloon. I’d even prepared a list of thermodynamic tasks enhanced by quantum phenomena: we can charge certain batteries at greater powers if we entangle them than if we don’t, entanglement can raise the amount of heat pumped out of a system by a refrigerator, etc. But Andrew’s question flummoxed me.

I bungled the answer. I toted out the aforementioned list, but it contained examples, not a unifying principle. The next day, I was sitting in an office borrowed from experimentalist Natalia Ares in New College, a Gothic confection founded during the late 1300s (as one should expect of a British college called “New”). Admiring the view of ancient stone walls, I realized how I should have responded the previous day.

View from a window near the office I borrowed in New College. If I could pack that office in a suitcase and carry it home, I would.

My answer begins with a blog post written in response to a quantum-thermodynamics question from a don at another venerable university: Yoram Alhassid. He asked, “What distinguishes quantum thermodynamics to quantum statistical mechanics?” You can read the full response here. Takeaways include thermodynamics’s operational flavor. When using an operational theory, we imagine agents who perform tasks, using given resources. For example, a thermodynamic agent may power a steamboat, given a hot gas and a cold gas. We calculate how effectively the agents can perform those tasks. For example, we compute heat engines’ efficiencies. If a thermodynamic agent can access quantum resources, I’ll call them “quantum thermodynamic.” If the agent can access only everyday resources, I’ll call them “classical thermodynamic.”

A quantum thermodynamic agent may access more resources than a classical thermodynamic agent can. The latter can leverage work (well-organized energy), free energy (the capacity to perform work), information, and more. A quantum agent may access not only those resources, but also entanglement (strong correlations between quantum particles), coherence (wavelike properties of quantum systems), squeezing (the ability to toy with quantum uncertainty as quantified by Heisenberg and others), and more. The quantum-thermodynamic agent may apply these resources as described in the list I rattled off at Andrew.

With Oxford experimentalist Natalia Ares in her lab

Yet quantum phenomena can impede a quantum agent in certain scenarios, despite assisting the agent in others. For example, coherence can reduce a quantum engine’s power. So can noncommutation. Everyday numbers commute under multiplication: 11 times 12 equals 12 times 11. Yet quantum physics features numbers that don’t commute so. This noncommutation underlies quantum uncertainty, quantum error correction, and much quantum thermodynamics blogged about ad nauseam on Quantum Frontiers. A quantum engine’s dynamics may involve noncommutation (technically, the Hamiltonian may contain terms that fail to commute with each other). This noncommutation—a fairly quantum phenomenon—can impede the engine similarly to friction. Furthermore, some quantum thermodynamic agents must fight decoherence, the leaking of quantum information from a quantum system into its environment. Decoherence needn’t worry any classical thermodynamic agent.

In short, quantum thermodynamic agents can benefit from more resources than classical thermodynamic agents can, but the quantum agents also face more threats. This principle might not encapsulate how all of quantum thermodynamics differs from its classical counterpart, but I think the principle summarizes much of the distinction. And at least I can posit such a principle. I didn’t have enough experience when I first authored a blog post about Oxford, in 2013. People say that Oxford never changes, but this quantum thermodynamic agent does.

In the University of Oxford Natural History Museum in 2013, 2017, and 2025. I’ve published nearly 150 Quantum Frontiers posts since taking the first photo!

1Oxford consists of colleges similarly to how neighborhoods form a suburb. Residents of multiple neighborhoods may work in the same dental office. Analogously, faculty from multiple colleges may work, and undergraduates from multiple colleges may major, in the same department.

Sean Carroll Thanksgiving

 (Apologies for the ugly blog format. We had a bit of a crash, and are working to get the template back in working order.)

This year we give thanks for a crucially important idea that can mean very different things to different people: information. (We’ve previously given thanks for the Standard Model LagrangianHubble’s Law, the Spin-Statistics Theoremconservation of momentumeffective field theorythe error bargauge symmetryLandauer’s Principle, the Fourier TransformRiemannian Geometrythe speed of lightthe Jarzynski equalitythe moons of Jupiterspaceblack hole entropyelectromagnetism, Arrow’s Impossibility Theorem, and quanta.)

“Information” is an idea that is everywhere in science and technology these days. From one angle it looks like such an obvious idea that it’s a bit startling to realize that information theory didn’t really come along until the work of Claude Shannon in the 1940s. From another, the idea has so many different shades of meaning that we shouldn’t be surprised (that’s a joke you will get in a bit) that it can be hard to understand.

Information theory is obviously an enormous subject, but we’re just giving thanks, not writing a textbook. I want to mention two ideas I find especially central. First, Shannon’s idea about relating information content to “surprisal.” Second, the very different intuitive notions of information that we get from engineering and physics.

Shannon, working at Bell Labs, was interested in the problem of how to send trustworthy signals efficiently over transatlantic cables. He was thinking about various ways to express information in a code: a set of symbols, each with a defined meaning. So a code might be an alphabet, or a set of words, or a literal cipher. And he noticed that there was a lot of redundancy in natural languages; the word “the” appears much more often in English than the word “axe,” although both have the same number of letters.

Let’s refer to each letter or symbol in a code as an “event.” Shannon’s insight was to realize that the more unlikely an event, the more information it conveyed when it was received. The statements “The Sun rose in the east this morning” and “The Sun rose in the west this morning” contain the same number of letters, but the former contains almost no information — you already were pretty sure the Sun would be rising in the east. But the latter, if obtained from a reliable source, would be very informative indeed, precisely because it was so unexpected. Clearly some kind of unprecedented astronomical catastrophe was in progress.

Imagine we can assign a probability p(x) to every different event x. Shannon wanted a way to quantify the information content of that event, which would satisfy various reasonable-seeming axioms: most crucially, that the information content of two independent events is the sum of the individual information contents. But the joint probability of two events is the product of their individual probabilities. So the natural thing to do would be to define the information content as the logarithm of the probability; the logarithm of a product equals the sum of the individual logarithms. But you want low probability to correspond to high information content, so Shannon defined the information content (also called the self-information, or surprisal, or Shannon information) of an event to be minus the log of the probability, which by math is equal to the log of the reciprocal of the probability:

    \[I(x) = - \log [p(x)] =\log \left(\frac{1}{p(x)}\right).\]

Note that probabilities are numbers between 0 and 1, and the log of such a number will be negative, with numbers closer to 0 being more negative than numbers closer to 1. So I(x) goes from +\infty at p(x)=0 to 0 at p(x)=1. An impossible message is infinitely surprising, and therefore conveys infinite information; an inevitable message is completely unsurprising, and conveys no information at all.

From there, Shannon suggested that we could characterize how efficient an entire code was at conveying information: just calculate the average (expectation value) of the information content for all possible events. When we have a probability distribution p(x), the average of any function f(x) is just the sum of the the values of the function times their respective probabilities, \langle f\rangle = \sum_x p(x) f(x). So we characterize the information content of a code via the quantity

    \[H[p] = - \sum_x p(x) \log[p(x)].\]

The only question is, what to call this lovely newly-defined quantity that surely nobody had ever thought of before? Happily Shannon was friends with John von Neumann, who informed him, “You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, no one really knows what entropy really is, so in a debate you will always have the advantage.” So entropy it is.

Indeed, this formula is precisely that which had been put forward (unknown to Shannon) by Josiah Willard Gibbs in the 1870’s as a definition of entropy in statistical mechanics. (It is related to the definition on Ludwig Boltzmann’s tombstone, S= k \log W, and Boltzmann had also suggested similar expressions to the above.) On the one hand, it seems remarkable to find precisely the same expression playing central roles in problems as disparate as sending signals across cables and watching cream mix into coffee; on the other hand, it’s a relatively simple expression and the axioms used to derive it are actually pretty similar, so perhaps we shouldn’t be surprised; on the third hand, the connection between information theory and statistical mechanics turns out to be deep and fruitful, so it’s more than just a mathematical coincidence.

But let me highlight the one aspect of the term “information” that can be sometimes confusing to people. To the engineer, a code that is maximally informative is one for which p(x) is relatively uniform over all events x, which means H[p(x)] is maximal or close to it; in that case, every event will tell you something at least a little bit interesting. For them, high entropy = high information.

But to a physicist who might be asking “how much information do I have about the state of a system?”, you have more information when p(x) is relatively narrowly concentrated around some value, rather than being all spread out. For them, high entropy = low information! Indeed, one physically-relevant notion of “information” is the “accessible information” of a system, which can be defined as H_\mathrm{max} - H. (I talk about this a bit in my recent solo podcast on complexity.)

Perhaps we shouldn’t be so surprised that physicists and engineers posit oppositely-directed relationships between entropy and information. It’s just a reflection of the fact that “information” is so ubiquitous and has so many different uses. We should be thankful that we’re beginning to understand it so well.

November 26, 2025

Tim GowersCreating a database of motivated proofs

It’s been over three years since my last post on this blog and I have sometimes been asked, understandably, whether the project I announced in my previous post was actually happening. The answer is yes — the grant I received from the Astera Institute has funded several PhD students and a couple of postdocs, and we have been busy. In my previous post I suggested that I would be open to remote collaboration, but that has happened much less, partly because a Polymath-style approach would have been difficult to manage while also ensuring that my PhD students would have work that they could call their own to put in their theses.

In general I don’t see a satisfactory solution to that problem, but in this post I want to mention a subproject of the main project that is very much intended to be a large public collaboration. A few months ago, a call came out from Renaissance Philanthropies saying that they were launching a $9m AI for Math Fund to spend on projects in the general sphere of AI and mathematics, and inviting proposals. One of the categories that they specifically mentioned was creating new databases, and my group submitted a proposal to create a database of what we call “structured motivated proofs,” a piece of terminology that I will explain a bit more later in just a moment. I am happy to report that our proposal was one of the 29 successful ones. Since a good outcome to the project will depend on collaboration from many people outside the group, we need to publicize it, which is precisely the purpose of this post. Below I will be more specific about the kind of help we are looking for.

Why might yet another database of theorems and proofs be useful?

The underlying thought behind this project is that AI for mathematics is being held back not so much by an insufficient quantity of data as by the wrong kind of data. (For a more general exploration of this theme, see here.) All mathematicians know, and some of us enjoy complaining about it, that it is common practice when presenting a proof in a mathematics paper, or even textbook, to hide the thought processes that led to the proof. Often this does not matter too much, because the thought processes may be standard ones that do not need to be spelt out to the intended audience. But when proofs start to get longer and more difficult, they can be hard to read because one has to absorb definitions and lemma statements that are not obviously useful, are presented as if they appeared from nowhere, and demonstrate their utility only much later in the argument.

A sign that this is a problem for AI is the behaviour one observes after asking an LLM to prove a statement that is too difficult for it. Very often, instead of admitting defeat, it will imitate the style of a typical mathematics paper and produce rabbits out of hats, together with arguments later on that those rabbits do the required job. The problem is that, unlike with a correct mathematics paper, one finds when one scrutinizes the arguments carefully that they are wrong. However, it is hard to find superficial features that distinguish between an incorrect rabbit with an incorrect argument justifying that rabbit (especially if the argument does not go into full detail) and a correct one, so the kinds of statistical methods used by LLMs do not have an easy way to penalize the incorrectness.

Of course, that does not mean that LLMs cannot do mathematics at all — they are remarkably good at it, at least compared with what I would have expected three years ago. How can that be, given the problem I have discussed in the previous paragraph?

The way I see it (which could change — things move so fast in this sphere), the data that is currently available to train LLMs and other systems is very suitable for a certain way of doing mathematics that I call guess and check. When trying to solve a maths problem, you will normally write down the routine parts of an argument without any fuss (and an LLM can do them too because it has seen plenty of similar examples), but if the problem as a whole is not routine, then at some point you have to stop and think, often because you need to construct an object that has certain properties (I mean this in a rather general way — the “object” might be a lemma that will split up the proof in a nice way) and it is not obvious how to do so. The guess-and-check approach to such moments is what it says: you make as intelligent a guess as you can and then see whether it has the properties you wanted. If it doesn’t, you make another guess, and you keep going until you get lucky.

The reason an LLM might be tempted to use this kind of approach is that the style of mathematical writing I described above makes it look as though that is what we as mathematicians do. Of course, we don’t actually do that, but we tend not to mention all the failed guesses we made and how we carefully examined why they failed, modifying them in appropriate ways in response, until we finally converged on an object that worked. We also don’t mention the reasoning that often takes place before we make the guess, saying to ourselves things like “Clearly an Abelian group can’t have that property, so I need to look for a non-Abelian group.”

Intelligent guess and check works well a lot of the time, particularly when carried out by an LLM that has seen many proofs of many theorems. I have often been surprised when I have asked an LLM a problem of the form \exists x\in X \ P(x), where P is some property that is hard to satisfy, and the LLM has had no trouble answering it. But somehow when this happens, the flavour of the answer given by the LLM leaves me with the impression that the technique it has used to construct x is one that it has seen before and regards as standard.

If the above picture of what LLMs can do is correct (the considerations for reinforcement-learning-based systems such as AlphaProof are not identical but I think that much of what I say in this post applies to them too for slightly different reasons), then the likely consequence is that if we pursue current approaches, then we will reach a plateau: broadly speaking they will be very good at answering a question if it is the kind of question that a mathematician with the right domain expertise and good instincts would find reasonably straightforward, but will struggle with anything that is not of that kind. In particular, they will struggle with research-level problems, which are, almost by definition, problems that experts in the area do not find straightforward. (Of course, there would probably be cases where an LLM spots relatively easy arguments that the experts had missed, but that wouldn’t fundamentally alter the fact that they weren’t really capable of doing research-level mathematics.)

But what if we had a database of theorems and proofs that did not hide the thought processes that lay behind the non-obvious details of the proofs? If we could train AI on a database of accounts of proof discoveries and if, having done so, we then asked it to provide similar accounts, then it would no longer resort to guess-and-check when it got stuck, because the proof-discovery accounts it had been trained on would not be resorting to it. There could be a problem getting it to unlearn its bad habits, but I don’t think that difficulty would be impossible to surmount.

The next question is what such a database might look like. One could just invite people to send in stream-of-consciousness accounts of how they themselves found certain proofs, but that option is unsatisfactory for several reasons.

  1. It can be very hard to remember where an idea came from, even a few seconds after one has had it — in that respect it is like a dream, the memory of which becomes rapidly less vivid as one wakes up.
  2. Often an idea will seem fairly obvious to one person but not to another.
  3. The phrase “motivated proof” means different things to different people, so without a lot of careful moderation and curation of entries, there is a risk that a database would be disorganized and not much more helpful than a database of conventionally written proofs.
  4. A stream-of-consciousness account could end up being a bit too much about the person who finds the proof and not enough about the mathematical reasons for the proof being feasibly discoverable.

To deal with these kinds of difficulties, we plan to introduce a notion of a structured motivated proof, by which we mean a proof that is generated in a very particular way that I will partially describe below. A major part of the project, and part of the reason we needed funding for it, is to create a platform that will make it convenient to input structured motivated proofs and difficult to insert the kinds of rabbits out of hats that make a proof mysterious and unmotivated. In this way we hope to gamify the task of creating the database, challenging people to input into our system proofs of certain theorems that appear to rely on “magic” ideas, and perhaps even offering prizes for proofs that contain steps that appear in advance to be particularly hard to motivate. (An example: the solution by Ellenberg and Gijswijt of the cap-set problem uses polynomials in a magic-seeming way. The idea of using polynomials came from an earlier paper of Croot, Lev and Pach that proved a closely related theorem, but in that paper it just appears in the statement of their Lemma 1, with no prior discussion apart from the words “in the present paper we use the polynomial method” in the introduction.)

What is a structured motivated proof?

I wrote about motivated proofs in my previous post, but thanks to many discussions with other members of the group, my ideas have developed quite a lot since then. Here are two ways we like to think about the concept.

1. A structured motivated proof is one that is generated by standard moves.

I will not go into full detail about what I mean by this, but will do so in a future post when we have created the platform that we would like people to use in order to input proofs into the database. But the basic idea is that at any one moment one is in a certain state, which we call a proof-discovery state, and there will be a set of possible moves that can take one from the current proof-discovery state to a new one.

A proof-discovery state is supposed to be a more formal representation of the state one is in when in the middle of solving a problem. Typically, if the problem is difficult, one will have asked a number of questions, and will be aware of logical relationships between them: for example, one might know that a positive answer to Q1 could be used to create a counterexample to Q2, or that Q3 is a special case of Q4, and so on. One will also have proved some results connected with the original question, and again these results will be related to each other and to the original problem in various ways that might be quite complicated: for example P1 might be a special case of Q2, which, if true would reduce Q3 to Q4, where Q3 is a generalization of the statement we are trying to prove.

Typically we will be focusing on one of the questions, and typically that question will take the form of some hypotheses and a target (the question being whether the hypotheses imply the target). One kind of move we might make is a standard logical move such as forwards or backwards reasoning: for example, if we have hypotheses of the form P(x) and \forall u\ P(u)\implies Q(u), then we might decide to deduce Q(x). But things get more interesting when we consider slightly less basic actions we might take. Here are three examples.

  1. We have in our list of hypotheses the fact that a function f is given by the formula f(x)=\exp(p(x)), where p is a polynomial, and our goal is to prove that there exists z such that f(z)=1. Without really thinking about it, we are conscious that f is a composition of two functions, one of which is continuous and one of which belongs to a class of functions that are all continuous, so f is continuous. Also, the conclusion \exists z\ f(z)=1 matches well the conclusion of the intermediate-value theorem. So the intermediate-value theorem comes naturally to mind and we add it to our list of available hypotheses. In practice we wouldn’t necessarily write it down, but the system we wish to develop is intended to model not just what we write down but also what is going on in our brains, so we propose a move that we call library extraction (closely related to what is often called premise selection in the literature). Note that we have to be a bit careful about library extraction. We don’t want the system to be allowed to call up results from the library that appear to be irrelevant but then magically turn out to be helpful, since those would feel like rabbits out of hats. So we want to allow extraction of results only if they are obvious given the context. It is not easy to define what “obvious” means, but there is a good rule of thumb for it: a library extraction is obvious if it is one of the first things ChatGPT thinks of when given a suitable non-cheating prompt. For example, I gave it the prompt, “I have a function f from the reals to the reals and I want to prove that there exists some z such that f(z)=1. Can you suggest any results that might be helpful?” and the intermediate-value theorem was its second suggestion. (Note that I had not even told it that f was continuous, so I did not need to make that particular observation before coming up with the prompt.)
  2. We have a goal of the form \exists x\in X\ P(x). If this were a Lean proof state, the most common way to discharge a goal of this form would be to input a choice for x. That is, we would instantiate the existential quantifier with some x_0 and our new goal would be P(x_0). However, as with library extraction, we have to be very careful about instantiation if we want our proof to be motivated, since we wish to disallow highly surprising choices of x_0 that can be found only after a long process of thought. So we have to restrict ourselves to obvious instantiations. One way that an instantiation in our system will count as obvious is if the variable is instantiated with a term that is already present in the proof-discovery state. If the desired term is not present, then in order to continue with the proof, it will be necessary to carry out moves that generate it. A very common technique for this is the use of metavariables: instead of guessing a suitable x_0, we create a variable x^\bullet and change the goal to P(x^\bullet), which we can think of as saying “I’m going to start trying to prove P(x^\bullet) even though I haven’t chosen x^\bullet yet. As the attempted proof proceeds, I will note down any properties Q_1,\dots,Q_k that x^\bullet might have that would help me finish the proof, in the hope that (i) I get to the end and (ii) the problem \exists x\ Q_1(x)\wedge\dots\wedge Q_k(x) is easier than the original problem.” Another kind of obvious instantiation is one where we try out an object that is “extreme” in some way — it might be the smallest element of X, or the largest, or the simplest. (Judging simplicity is another place where the ChatGPT rule of thumb can be used.)
  3. We cannot see how to answer the question we are focusing on so we ask a related question. Two very common kinds of related question (as emphasized by Polya) are generalization and specialization. Perhaps we don’t see why a hypothesis is helpful, so we see whether the result holds if we drop that hypothesis. If it does, then we are no longer distracted by an irrelevant hypothesis. If it does not, then we can hope to find a counterexample that will help us understand how to use the hypothesis. Or perhaps we are trying to prove a general statement but it is not clear how to do so, so instead we formulate some special cases, hoping that we can prove them and spot features of the proofs that we can generalize. Again we have to be rather careful here not to allow “non-obvious” generalizations and specializations. Roughly the idea there is that a generalization should be purely logical — for example, dropping a hypothesis is fine but replacing the hypothesis “f is twice differentiable” by “f is upper semicontinuous” is not — and that a specialization should be to a special case that counts as an obvious instantiation in the sense discussed just above.

2. A structured motivated proof is one that can be generated with the help of a point-and-click system.

This is a surprisingly useful way to conceive of what we are talking about, especially as it relates closely to what I was talking about earlier: imposing a standard form on motivated proofs (which is why we call them “structured” motivated proofs) and gamifying the process of producing them.

The idea is that a structured motivated proof is one that can be generated using an interface (which we are in the process of creating — at the moment we have a very basic prototype that has a few of the features we will need, but not yet the more interesting ones) that has one essential property: the user cannot type in data. So what can they do? They can select text that is on their screen (typically mathematical expressions or subexpressions), they can click buttons, choose items from drop-down menus, and accept or reject “obvious” suggestions made to them by the interface.

If, for example, the current goal is an existential statement \exists x\ P(x), then typing in a formula that defines a suitable x is not possible, so instead one must select text or generate new text by clicking buttons, choosing from short drop-down menus, and so on. This forces the user to generate x, which is our proxy for showing where the idea of using x came from.

Broadly speaking, the way the prototype works is to get an LLM to read a JSON object that describes the variables, hypotheses and goals involved in the proof state in a structured format, and to describe (by means of a fairly long prompt) the various moves it might be called upon to do. Thus, the proofs generated by the system are not formally verified, but that is not an issue that concerns us in practice since there will be a human in the loop throughout to catch any mistakes that the LLM might make, and this flexibility may even work to our advantage to better capture the fluidity of natural-language mathematics.

There is obviously a lot more to say about what the proof-generating moves are, or (approximately equivalently) what the options provided by a point-and-click system will be. I plan to discuss that in much more detail when we are closer to having an interface ready, the target for which is the end of this calendar year. But the aim of the project is to create a database of examples of proofs that have been successfully generated using the interface, which can then be used to train AI to play the generate-structured-motivated-proof game.

How to get involved.

There are several tasks that will need doing once the project gets properly under way. Here are some of the likely ones.

  1. The most important is for people to submit structured motivated (or move-generated) proofs to us on the platform we provide. We hope that the database will end up containing proofs of a wide range of difficulty (of two kinds — there might be fairly easy arguments that are hard to motivate and there might be arguments that are harder to follow but easier to motivate) and also a wide range of areas of mathematics. Our initial target, which is quite ambitious, is to have around 1000 entries by two years from now. While we are not in a position to accept entries yet, if you are interested in participating, then it is not too early to start thinking in a less formal way about how to convert some of your favourite proofs into motivated versions, since that will undoubtedly make it easier to get them accepted by our platform when it is ready.
  2. We are in the process of designing the platform. As I mentioned earlier, we already have a prototype, but there are many moves we will need it to be able to do that it cannot currently do. For example, the current prototype allows just a single proof state, which consists of some variable declarations, hypotheses, and goals. It does not yet support creating subsidiary proof states (which we would need if we wanted to allow the user to consider generalizations and specializations, for example). Also, for the moment the prototype gets an LLM to implement all moves, but some of the moves, such as applying modus ponens, are extremely mechanical and would be better done using a conventional program. (On the other hand, moves such as “obvious library extraction” or “provide the simplest example” are better done by an LLM.) Thirdly, a technical problem is that LaTeX is currently rendered as images, which makes it hard to select subexpressions, something we will need to be able to do in a non-clunky way. And the public version of the platform will need to be web-based and very convenient to use. We will want features such as being able to zoom out and look at some kind of dependency diagram of all the statements and questions currently in play, and then zoom in on various nodes if the user wishes to work on them. If you think you may be able (and willing) to help with some of these aspects of the platform, then we would be very happy to hear from you. For some, it would probably help to have a familiarity with proof assistants, while for others we would be looking for somebody with software engineering experience. The grant from the AI for Math Fund will allow us to pay for some of this help, at rates to be negotiated. We are not yet ready to specify in detail what help we need, but would welcome any initial expressions of interest.
  3. Once the platform is ready and people start to submit proofs, it is likely that, at least to start with, they will find that the platform does not always provide the moves they need. Perhaps they will have a very convincing account of where a non-obvious idea in the proof came from, but the system won’t be expressive enough for them to translate that account into a sequence of proof-generating moves. We will want to be able to react to such situations (if we agree that a new move is needed) by expanding the capacity of the platform. It will therefore be very helpful if people sign up to be beta-testers, so that we can try to get the platform to a reasonably stable state before opening it up to a wider public. Of course, to be a beta-tester you would need to have a few motivated proofs in mind.
  4. It is not obvious that every proof submitted via the platform, even if submitted successfully, would be a useful addition to the database. For instance, it might be such a routine argument that no idea really needs to have its origin explained. Or it might be that, despite our best efforts, somebody finds a way of sneaking in a rabbit while using only the moves that we have provided. (One way this could happen is if an LLM made a highly non-obvious suggestion that happened to work, in which case the rule of thumb that if an LLM thinks of it, it must be obvious, would have failed in that instance.) For this reason, we envisage having a team of moderators, who will check entries and make sure that they are good additions to the database. We hope that this will be an enjoyable task, but it may have its tedious aspects, so we envisage paying moderators — again, this expense was allowed for in our proposal to the AI for Math Fund.

If you think you might be interested in any of these roles, please feel free to get in touch. Probably the hardest recruitment task for us will be identifying the right people with the right mixture of mathematical knowledge and software engineering skills to help us turn the platform into a well-designed web-based one that is convenient and pleasurable to use. If you think you might be such a person, or if you have a good idea for how we should go about finding one, we would be particularly interested to hear from you.

In a future post, I will say more about the kinds of moves that our platform will allow, and will give examples of non-motivated proofs together with how motivated versions of those proofs can be found and entered using the platform (which may involve a certain amount of speculation about what the platform will end up looking like).

How does this relate to use of tactics in a proof assistant?

In one way, our “moves” can be regarded as tactics of a kind. However, some of the moves we will need are difficult to implement in conventional proof assistants such as Lean. In parallel with the work described above, we hope to create an interface to Lean that would allow one to carry out proof-discovery moves of the kind discussed above but with the proof-discovery states being collections of Lean proof states. Members of my group have already been working on this and have made some very interesting progress, but there is some way to go. However, we hope that at some point (and this is also part of the project pitched to the AI for Math Fund) we will have created another interface that will have Lean working in the background, so that it will be possible to generate motivated proofs that will be (or perhaps it is better to say include) proofs in Lean at the same time.

Another possibility that we are also considering is to use the output of the first platform (which, as mentioned above, will be fairly formal, but not in the strict sense of a language such as Lean) to create a kind of blueprint that can then be autoformalized automatically. Then we would have a platform that would in principle allow mathematicians to search for proofs while working on their computers without having to learn a formal language, with their thoughts being formalized as they go.

November 25, 2025

David Hoggsubstellar objects (brown dwarfs)

I spent the day at the NSBP / NSHP meeting in San José. My favorite session of the day was the morning astro session, which was entirely about brown dwarfs. I learned a lot in a very short time. Caprice Phillips (UCSC) introduced the session with an introduction to the scientific and technical questions in play. She put a lot of emphasis on using binaries and clusters to put detailed abundance ratios onto substellar objects. This was what I expected: I thought (walking in to this session) that all known abundance ratios for brown dwarfs were from such kinds of studies. I learned different (keep reading).

Gabriel Munoz Zarazua (SFSU) followed by showing spectra from M-dwarfs, brown dwarfs, and Jupiter. It definitely looks like a sequence. He does spectral fitting (what they call, in this business, retrievals). It looks like he is getting very good, somewhat precise, abundance ratios for the photospheres of substellar objects! I asked more about this in the question period, and apparently I am way behind the times (Emily Rauscher, Michigan, helpfully pointed this out to me): Now brown-dwarf photosphere models are so good, they can be used to measure abundances, and pretty well.

I also learned in this session (maybe from Jorge Sanchez, ASU, or maybe from Efrain Alvarado, SFSU) that there is a very strong mass–abundance relation in the Solar System. That is, we don't expect, if brown dwarfs form the way planets do, that the detailed abundances of the brown dwarfs will match exactly the detailed abundances of the primary stars. But now we are really in a position to test that. Sanchez showed that we can get, from even photometry, abundances for substellar objects in the Milky Way halo. Again, totally new to me! And he finds metallicities at or below −3. Alvarado showed data on an amazing system J1416, which is an L–T binary with no stellar companion. Apparently it is the only known completely substellar binary.

Tommaso DorigoBaby Steps In The Reinforcement Learning World

I am moving some baby steps in the direction of Reinforcement Learning (RL) these days. In machine learning, RL is a well-established and very promising avenue for the development of artificial intelligence, and the field is in rapid development. Unfortunately I have been left behind, as I never really needed to fiddle with those techniques for my research. Until recently.

read more

November 23, 2025

Scott Aaronson Podcasts!

A 9-year-old named Kai (“The Quantum Kid”) and his mother interviewed me about closed timelike curves, wormholes, Deutsch’s resolution of the Grandfather Paradox, and the implications of time travel for computational complexity:

This is actually one of my better podcasts (and only 24 minutes long), so check it out!


Here’s a podcast I did a few months ago with “632nm” about P versus NP and my other usual topics:


For those who still can’t get enough, here’s an interview about AI alignment for the “Hidden Layers” podcast that I did a year ago, and that I think I forgot to share on this blog at the time:


What else is in the back-catalog? Ah yes: the BBC interviewed me about quantum computing for a segment on Moore’s Law.


As you may have heard, Steven Pinker recently wrote a fantastic popular book about the concept of common knowledge, entitled When Everyone Knows That Everyone Knows… Steve’s efforts render largely obsolete my 2015 blog post Common Knowledge and Aumann’s Agreement Theorem, one of the most popular posts in this blog’s history. But I’m willing to live with that, not only because Steven Pinker is Steven Pinker, but also because he used my post as a central source for the topic. Indeed, you should watch his podcast with Richard Hanania, where Steve lucidly explains Aumann’s Agreement Theorem, noting how he first learned about it from this blog.

November 22, 2025

John BaezBeyond the Geometry of Music

Yesterday I had a great conversation with Dmitri Tymoczko about groupoids in music theory. But at this Higgs Centre Colloquium, he preferred to downplay groupoids and talk in a way physicists would enjoy more. Click on the picture to watch his talk!

What’s great is that he’s not faking it: he’s really found deep ways in which symmetry shows up pervasively in music.

At first he tried to describe them geometrically using ‘orbifolds’, which are spaces in which some singular points have nontrivial symmetry groups, like the tip of a cone. But then he realized that the geometry was less important than the symmetry, which you can describe using ‘groupoids’: categories where every morphism is invertible. That’s why his talk is called “Beyond the geometry of music”.

I’m helping him with his work on groupoids, and I hope he explains his work to mathematicians someday without pulling his punches. I didn’t get to interview him yesterday, but I’ll try to do that soon.

For now you can read his books A Geometry of Music and Harmony: an Owner’s Manual along with many papers. What I’ve read so far is really exciting.

November 21, 2025

Doug NatelsonQuantum geometry - some intuition

There has been a great growing interest in quantum geometry in recent years.  Last week, I heard an excellent talk by Raquel Queiroz about this that gave me a more physically intuitive interpretation  of this topic.  The more formal write-up is in this preprint from this past April, which I'd missed at the time.

Caution:  Math incoming.  I will try to give a more physical picture at the end.  I know that this won't be very readable to non-experts.    

As I've written before,  (e.g. here and a bit here), the electronic states in crystalline solids are often written as Bloch waves of the form \(u_{n\mathbf{k}}(\mathbf{r})\exp(i \mathbf{k}\cdot \mathbf{r})\), where \(u_{n\mathbf{k}}(\mathbf{r})\) is periodic in the spatial period of the crystal lattice.  For many years, the \(\mathbf{k}\) dependence of \(u_{n\mathbf{k}}(\mathbf{r})\) was comparatively neglected, but now it is broadly appreciated that this is the root of all kinds of interesting physics, including the anomalous Hall effect and its quantum version.  

We can compute how much \(u_{n\mathbf{k}}(\mathbf{r})\) changes with \(\mathbf{k}\).  The Berry connection is related to the phase angle racked up by moving around in \(\mathbf{k}\), and it's given by \( \mathbf{A}(\mathbf{k}) = i \langle u_{n\mathbf{k}}| \nabla_{\mathbf{k}}| u_{n\mathbf{k}} \rangle \).  One can define \(\mathbf{\Omega} \equiv \nabla \times \mathbf{A}(\mathbf{k})\) as the Berry curvature, and the "anomalous velocity" is given by \(-\dot{\mathbf{k}}\times \mathbf{\Omega}\).  

If we worry about possible changes in the magnitude as well, and \( |\langle u_{n\mathbf{k}}| u_{n\mathbf{k+dk}} \rangle |^{2} = 1 - g^{n}_{\mu \nu}dk_{\mu}dk_{\nu}\) plus higher order terms.  The quantity \(g^{n}_{\mu \nu}\) is the quantum metric, and it can be written in terms of dipole operators:  \(g^{n}_{\mu \nu}= \sum_{m\ne n}\langle u_{n,\mathbf{k}}|\hat{r}_{\mu}|u_{m \mathbf{k}}\rangle \langle u_{m,\mathbf{k}}|\hat{r}_{\nu}|u_{n \mathbf{k}}\rangle\).  The quantum metric quantifies the "distance between" the Bloch states as one moves around in \(\mathbf{k}\).  

That last bit is what I really learned from the talk.  Basically, if you try to consider electrons localized to a particular lattice site in real space, this can require figuring in states in multiple bands, and the matrix elements involve dipole operators.  The quantum geometric tensor \(g_{\mu \nu}\) quantifies the dipole fluctuations in the electronic density.  You can define a lengthscale \(\ell_{g}\equiv \sqrt{\mathrm{Tr} g}\), and this can tell you about the spatial scale of polarization fluctuations relative to, e.g., the lattice spacing.  Metals will have essentially divergent fluctuation lengthscales, while insulators have nicely bound charges (that give peaks in the optical conductivity at finite frequency).   The quantum geometry then influences all kinds of experimentally measurable quantities (see here).  

Neat stuff.  Someday I'd like to return to this with a nice cartoon/animation/presentation for non-experts.  The idea that there is so much richness within even relatively "boring" materials still amazes me.


Matt von HippelMandatory Dumb Acronyms

Sometimes, the world is silly for honest, happy reasons. And sometimes, it’s silly for reasons you never even considered.

Scientific projects often have acronyms, some of which are…clever, let’s say. Astronomers are famous for acronyms. Read this list, and you can find examples from 2D-FRUTTI and ABRACADABRA to WOMBAT and YORIC. Some of these aren’t even “really” acronyms, using letters other than the beginning of each word, multiple letters from a word, or both. (An egregious example from that list: VESTALE from “unVEil the darknesS of The gAlactic buLgE”.)

But here’s a pattern you’ve probably not noticed. I suggest that you should see more of these…clever…acronyms in projects in Europe, and they should show up in a wider range of fields, not just astronomy. And the reason why, is the European Research Council.

In the US, scientific grants are spread out among different government agencies. Typical grants are small, the kind of thing that lets a group share a postdoc every few years, with different types of grants covering projects of different scales.

The EU, instead, has the European Research Council, or ERC, with a flagship series of grants covering different career stages: Starting, Consolidator, and Advanced. Unlike most US grants, these are large (supporting multiple employees over several years), individual (awarded to a single principal investigator, not a collaboration) and general (the ERC uses the same framework across multiple fields, from physics to medicine to history).

That means there are a lot of medium-sized research projects in Europe that are funded by an ERC grant. And each of them are required to have an acronym.

Why? Who knows? “Acronym” is simply one of the un-skippable entries in the application forms, with a pre-set place of honor in their required grant proposal format. Nobody checks whether it’s a “real acronym”, so in practice it often isn’t, turning into some sort of catchy short name with “acronym vibes”. It, like everything else on these forms, is optimized to catch the attention of a committee of scientists who really would rather be doing something else, often discussed and refined by applicants’ mentors and sometimes even dedicated university staff.

So if you run into a scientist in Europe who proudly leads a group with a cutesy, vaguely acronym-adjacent name? And you keep running into these people?

It’s not a coincidence, and it’s not just scientists’ sense of humor. It’s the ERC.

Scott Aaronson Quantum Investment Bros: Have you no shame?

Near the end of my last post, I made a little offhand remark:

[G]iven the current staggering rate of hardware progress, I now think it’s a live possibility that we’ll have a fault-tolerant quantum computer running Shor’s algorithm before the next US presidential election. And I say that not only because of the possibility of the next US presidential election getting cancelled, or preempted by runaway superintelligence!

As I later clarified, I’ll consider this “live possibility” to be fulfilled even if a fault-tolerant Shor’s algorithm is “merely” used to factor 15 into 3×5—a milestone that seems a few steps, but only a few steps, away from what Google, Quantinuum, QuEra, and others have already demonstrated over the past year. After that milestone, I then expect “smooth sailing” to more and more logical qubits and gates and the factorization of larger and larger integers, however fast or slow that ramp-up proceeds (which of course I don’t know).

In any case, the main reason I made my remark was just to tee up the wisecrack about whether I’m not sure if there’ll be a 2028 US presidential election.


My remark, alas, then went viral on Twitter, with people posting countless takes like this:

A quantum expert skeptic who the bears quote all the time – Scott Aaronson – recently got very excited about a number of quantum advances. He now thinks there’s a possibility of running Shor before the next US president election – a timeline that lines up ONLY with $IONQ‘s roadmap, and NOBODY else’s! This represent a MAJOR capitulation of previously predicted timelines by any skeptics.

Shall we enumerate the layers of ugh here?

  1. I’ve been saying for several years now that anyone paranoid about cybersecurity should probably already be looking to migrate to quantum-resistant cryptography, because one can’t rule out the possibility that hardware progress will be fast. I didn’t “capitulate”: I mildly updated what I said before, in light of exciting recent advances.
  2. A “live possibility” is short not only of a “certainty,” but of a “probability.” It’s basically just an “I’m not confident this won’t happen.”
  3. Worst is the obsessive focus on IonQ, a company that I never mentioned (except in the context of its recently-acquired subsidiary, Oxford Ionics), but which now has a $17 billion valuation. I should explain that, at least since it decided to do an IPO, IonQ has generally been regarded within the research community as … err … a bit like the early D-Wave, intellectual-respectability-wise. They’ll eagerly sell retail investors on the use of quantum computers to recognize handwriting and suchlike, despite (I would say) virtually no basis to believe in a quantum scaling advantage for such tasks. Or they’ll aggressively market current devices to governments who don’t understand what they’re for, but just want to say they have a quantum computer and not get left behind. Or they’ll testify to Congress that quantum, unlike AI, “doesn’t hallucinate” and indeed is “deterministic.” It pains me to write this, as IonQ was founded by (and indeed, still employs) scientists who I deeply admire and respect.
  4. Perhaps none of this would matter (or would matter only to pointy-headed theorists like me) if IonQ were the world leader in quantum computing hardware, or even trapped-ion hardware. But by all accounts, IonQ’s hardware and demonstrations have lagged well behind those of its direct competitor, Quantinuum. It seems to me that, to whatever extent IonQ gets vastly more attention, it’s mostly just because it chose to IPO early, and also because it’s prioritized marketing to the degree it has.

Over the past few days, I’ve explained the above to various people, only to have them look back at me with glazed, uncomprehending eyes and say, “so then, which quantum stock should I buy? or should I short quantum?”

It would seem rude for me to press quarters into these people’s hands, explaining that they must make gain from whatever they learn. So instead I reply: “You do realize, don’t you, that I’m, like, a professor at a state university, who flies coach and lives in a nice but unremarkable house? If I had any skill at timing the market, picking winners, etc., don’t you think I’d live in a mansion with an infinity pool, and fly my Cessna to whichever conferences I deigned to attend?”


It’s like this: if you think quantum computers able to break 2048-bit cryptography within 3-5 years are a near-certainty, then I’d say your confidence is unwarranted. If you think such quantum computers, once built, will also quickly revolutionize optimization and machine learning and finance and countless other domains beyond quantum simulation and cryptanalysis—then I’d say that more likely than not, an unscrupulous person has lied to you about our current understanding of quantum algorithms.

On the other hand, if you think Bitcoin, and SSL, and all the other protocols based on Shor-breakable cryptography, are almost certainly safe for the next 5 years … then I submit that your confidence is also unwarranted. Your confidence might then be like most physicists’ confidence in 1938 that nuclear weapons were decades away, or like my own confidence in 2015 that an AI able to pass a reasonable Turing Test was decades away. It might merely be the confidence that “this still looks like the work of decades—unless someone were to gather together all the scientific building blocks that have now been demonstrated, and scale them up like a stark raving madman.” The trouble is that sometimes people, y’know, do that.

Beyond that, the question of “how many years?” doesn’t even interest me very much, except insofar as I can mine from it the things I value in life, like scientific understanding, humor, and irony.


There are, famously, many intellectual Communists who are ruthless capitalists in their day-to-day lives. I somehow wound up the opposite. Intellectually, I see capitalism as a golden goose, a miraculous engine that’s lifted the human species out of its disease-ridden hovels and into air-conditioned high-rises, whereas Communism led instead to misery and gulags and piles of skulls every single time it was tried.

And yet, when I actually see the workings of capitalism up close, I often want to retch. In case after case, it seems, our system rewards bold, confident, risk-taking ignoramuses and liars, those who can shamelessly hype a technology (or conversely, declare it flatly impossible)—with such voices drowning out the cautious experts who not only strive to tell the truth, but also made all the actual discoveries that the technology rests on. My ideal economic system is, basically, whichever one can keep the people who can clearly explain the capabilities and limits and risks and benefits of X in charge of X for as long as possible.

November 19, 2025

John BaezSafeguarded AI (Part 2)

60 people, including a lot of category theorists, are meeting in Edinburgh for the £59 million UK project called Safeguarded AI. I talked about it before here.

The plan is to build software that will let you precisely specify systems of many kinds, which an AI will design, and verify that what the AI designed meets your specifications. So: it’s not about building an AI, but instead, building a way to specify jobs for it and verify that it did those jobs correctly!

The director of this project, David Dalrymple, has changed the plan recently. There were many teams of category theorists designing formalisms to get this job done. David Jaz Myers at Topos Research UK was supposed to integrate all these formalisms. That would be a huge job.

But recently all but a few teams have been cut off from the main project—they can now do whatever they want. The project will focus on 3 parts:

1) The “categorical core”: a software infrastructure that lets you program using category theory concepts. I think Amar Hadzihasanovic, my former student Owen Lynch, and two others will be building this.

2) “DOTS”: the double operadic theory of systems, a general framework for building systems out of smaller parts. This is David Jaz Myers’ baby—see the videos.

3) Example applications. One of these, building colored Petri nets, will be done by my former student Jade Master. I don’t know all the others.

By September 2026, David Jaz Myers, Sophie Libkind, Matteo Capucci, Jason Brown and others are supposed to write a 300-page “thesis” on how this whole setup works. Some of the ideas are already available here:

• David Jaz Myers and Sophie Libkind, Towards a double operadic theory of systems.

It feels funny that so much of the math I helped invent is going into this project, and there’s a massive week-long meeting about it just a ten minute walk away, but I’m not involved. But this was by choice, and I’m happier just watching.

I apologize for any errors in the above, and for leaving out many other names of people who must be important in this project. I’ve spoken to various people involved, but not enough. I’m going to talk to David Jaz Myers tomorrow, but he wants to talk about what I’m really interested in these days: octonions and particle physics!

November 17, 2025

John BaezThe Inverse Cube Force Law

Newton’s Principia is famous for its investigations of the inverse square force law for gravity. But in this book Newton also did something that remained little-known until fairly recently. He figured out what kind of central force exerted upon a particle can rescale its angular velocity by a constant factor without affecting its radial motion. This turns out to be a force obeying an inverse cube law.

Given a particle in Euclidean space, a central force is a force that points toward or away from the origin and depends only on the particle’s distance from the origin. If the particle’s position at time t is \mathbf{r}(t) \in \mathbb{R}^n and its mass is some number m > 0, we have

m \, \ddot{\mathbf{r}}(t) = F(r(t))  \,\hat{\mathbf{r}}(t)

where \hat{\mathbf{r}}(t) is a unit vector pointing outward from the origin at the point \mathbf{r}(t). A particle obeying this equation always moves in a plane through the origin, so we can use polar coordinates and write the particle’s position as \bigl(r(t), \theta(t)\bigr). With some calculation one can show the particle’s distance from the origin, r(t), obeys

\displaystyle{ m \ddot r(t) = F(r(t)) + \frac{L^2}{mr(t)^3}  \qquad \qquad  \qquad \qquad (1) }

Here L = mr(t)^2 \dot \theta(t), the particle’s angular momentum, is constant in time. The second term in equation (1) says that the particle’s distance from the origin changes as if there were an additional force pushing it outward. This is a “fictitious force”, an artifact of working in polar coordinates. It is called the centrifugal force. And it obeys an inverse cube force law!

This explains Newton’s observation. Let us see why. Suppose that we have two particles moving in two different central forces F_1 and F_2, each obeying a version of equation (1), with the same mass m and the same radial motion r(t), but different angular momenta L_1 and L_2. Then we must have

\displaystyle{      F_1(r(t)) + \frac{L_1^2}{mr(t)^3}  =  F_2(r(t)) + \frac{L_2^2}{mr(t)^3}  }

If the particle’s angular velocities are proportional then L_2 = kL_1 for some constant k, so

\displaystyle{       F_2(r_1(t)) - F_1(r(t)) = \frac{(k^2 - 1)L_1^2}{mr(t)^3} }

This says that F_2 equals F_1 plus an additional inverse cube force.

A particle’s motion in an inverse cube force has curious features. First compare Newtonian gravity, which is an attractive inverse square force, say F(r) = -c/r^2 with c > 0. In this case we have

\displaystyle{ m \ddot r(t) = -\frac{c}{r(t)^2} + \frac{L^2}{mr(t)^3 } }

Because 1/r^3 grows faster than 1/r^2 as r \downarrow 0, as long as the angular momentum L is nonzero the repulsion of the centrifugal force will beat the attraction of gravity for sufficiently small r, and the particle will not fall in to the origin. The same is true for any attractive force F(r) = -c/r^p with p < 3. But an attractive inverse cube force can overcome the centrifugal force and make a particle fall in to the origin.

In fact there are three qualitatively different possibilities for the motion of a particle in an attractive inverse cube force F(r) = -c/r^3, depending on the value of c. With work we can solve for 1/r as a function of \theta (which is easier than solving for r). There are three cases depending on the value of

\displaystyle{  \omega^2 = 1 - \frac{cm}{L^2} }

vaguely analogous to the elliptical, parabolic and hyperbolic orbits of a particle in an inverse square force law:

\displaystyle{ \frac{1}{r(\theta)} } = \left\{ \begin{array}{lcl}   A \cos(\omega \theta) + B \sin(\omega \theta) & \text{if} & \omega^2 > 0 \\ \\   A + B \theta & \text{if} & \omega = 0 \\  \\   A e^{|\omega| \theta} + B e^{-|\omega| \theta}  & \text{if} & \omega^2 < 0   \end{array} \right.

The third case occurs when the attractive inverse cube force is strong enough to overcome the centrifugal force: c > L^2/m. Then the particle can spiral in to its doom, hitting the origin in a finite amount of time after infinitely many orbits, like this:

All three curves above are called Cotes spirals, after Roger Cotes’ work on the inverse cube force law, published posthumously in 1722. Cotes seems to have been the first to compute the derivative of the sine function. After Cotes’ death at the age of 33, Newton supposedly said “If he had lived we would have known something.”

The subtlety of the inverse cube force law is greatly heightened when we study it using quantum rather than classical mechanics. Here if c is too large the theory is ill-defined, because there is no reasonable choice of self-adjoint Hamiltonian. If c is smaller the theory is well-behaved. But at a certain borderline point it exhibits a remarkable property: spontaneous breaking of scaling symmetry. I hope to discuss this in my next column.

For more on the inverse cube force law, see:

• N. Grossman, The Sheer Joy of Celestial Mechanics, Birkhäuser, Basel, 1996, p. 34.

For more on Newton’s work involving the inverse cube force law, see:

• Wikipedia, Newton’s theorem of revolving orbits.

• S. Chandrasekhar, Newton’s Principia for the Common Reader, Oxford U. Press, Oxford, 1995, pp. 183–200.

Cotes’ book is

• Roger Cotes, Harmonia Mensuarum, Cambridge, 1722.

November 14, 2025

Scott Aaronson Quantum computing: too much to handle!

Tomorrow I’m headed to Berkeley for the Inkhaven blogging residency, whose participants need to write one blog post per day or get kicked out. I’ll be there to share my “wisdom” as a distinguished elder blogger (note that Shtetl-Optimized is now in its twentieth year). I’m acutely aware of the irony, that I myself can barely muster the willpower these days to put up a post every other week.

And it’s not as if nothing is happening in this blog’s traditional stomping-ground of quantum computing! In fact, the issue is just the opposite: way too much is happening for me to do it any sort of justice. Who do people think I am, Zvi Mowshowitz? The mere thought of being comprehensive, of responsibly staying on top of all the latest QC developments, makes me want to curl up in bed, and either scroll through political Substacks or take a nap.


But then, you know, eventually a post gets written. Let me give you some vignettes about what’s new in QC, any one of which could easily have been its own post if I were twenty years younger.

(1) Google announced verifiable quantum advantage based on Out-of-Time-Order-Correlators (OTOC)—this is actually from back in June, but it’s gotten more and more attention as Google has explained it more thoroughly. See especially this recent 2-page note by King, Kothari, et al., explaining Google’s experiment in theoretical computer science language. Basically, what they do is, starting from the all-|0⟩ state, to apply a random circuit C, then a single gate g, then C-1, then another gate h, then C again, then g again, then C-1, and then measure a qubit. If C is shallow, then the qubit is likely to still be |0⟩. If C is too deep, then the qubit is likely to be in the maximally mixed state, totally uncorrelated with its initial state—the gates g and h having caused a “butterfly effect” that completely ruined all the cancellation between C and C-1. Google claims that, empirically, there’s an intermediate regime where the qubit is neither |0⟩ nor the maximally mixed state, but a third thing—and that this third thing seems hard to determine classically, using tensor network algorithms or anything else they’ve thrown at it, but it can of course be determined by running the quantum computer. Crucially, because we’re just trying to estimate a few parameters here, rather than sample from a probability distribution (as with previous quantum supremacy experiments), the output can be checked by comparing it against the output of a second quantum computer, even though the problem still isn’t in NP. Incidentally, if you’re wondering why they go back and forth between C and C-1 multiple times rather than just once, it’s to be extra confident that there’s not a fast classical simulation. Of course there might turn out to be a fast classical simulation anyway, but if so, it will require a new idea: gauntlet thrown.

(2) Quantinuum, the trapped-ion QC startup in Colorado, announced its Helios processor. Quick summary of the specs: 98 qubits, all-to-all 2-qubit gates with 99.92% fidelity, the ability to choose which gates to apply “just in time” (rather than fixing the whole circuit in advance, as was needed with their previous API), and an “X”-shaped junction for routing qubits one way or the other (the sort of thing that a scalable trapped-ion quantum computer will need many of). This will enable, and is already enabling, more and better demonstrations of quantum advantage.

(3) Quantinuum and JP Morgan Chase announced the demonstration of a substantially improved version of my and Shih-Han-Hung’s protocol for generating cryptographically certified random bits, using quantum supremacy experiments based on random circuit sampling. They did their demo on Quantinuum’s new Helios processor. Compared to the previous demonstration, the new innovation is to send the circuit to the quantum computer one layer at a time, rather than all at once (something that, again, Quantinuum’s new API allows). The idea is that a cheating server, who wanted to spoof the randomness deterministically, now has much less time: using the most competitive known methods (e.g., those based on tensor network contraction), it seems the cheater would need to swing into action only after learning the final layer of gates, so would now have mere milliseconds to spoof rather than seconds, making Internet latency the dominant source of spoofing time in practice. While a complexity-theoretic analysis of the new protocol (or, in general, of “layer-by-layer” quantum supremacy protocols like it) is still lacking, I like the idea a lot.

(4) The startup company BlueQubit announced a candidate demonstration of verifiable quantum supremacy via obfuscated peaked random circuits, again on a Quantinuum trapped-ion processor (though not Helios). In so doing, BlueQubit is following the program that Yuxuan Zhang and I laid out last year: namely, generate a quantum circuit C that hopefully looks random to any efficient classical algorithm, but that conceals a secret high-probability output string x, which pops out if you run C on a quantum computer on the all-0 initial state. To try to hide x, BlueQubit uses at least three different circuit obfuscation techniques, which already tells you that they can’t have complete confidence in any one of them (since if they did, why the other two?). Nevertheless, I’m satisfied that they tried hard to break their own obfuscation, and failed. Now it’s other people’s turn to try.

(5) Deshpande, Fefferman, et al. announced a different theoretical proposal for quantum advantage from peaked quantum circuits, based on error-correcting codes. This seems tempting to try to demonstrate along the way to quantum fault-tolerance.

(6) A big one: John Bostanci, Jonas Haferkamp, Chinmay Nirkhe, and Mark Zhandry announced a proof of a classical oracle separation between the complexity classes QMA and QCMA, something that they’ve been working on for well over a year. Their candidate problem is basically a QMA-ified version of my Forrelation, which Raz and Tal previously used to achieve an oracle separation between BQP and PH. I caution that their paper is 91 pages long and hasn’t yet been vetted by independent experts, and there have been serious failed attempts on this exact problem in this past. If this stands, however, it finally settles a problem that’s been open since 2002 (and which I’ve worked on at various points starting in 2002), and shows a strong sense in which quantum proofs are more powerful than classical proofs. Note that in 2006, Greg Kuperberg and I gave a quantum oracle separation between QMA and QCMA—introducing the concept of quantum oracles for the specific purpose of that result—and since then, there’s been progress on making the oracle steadily “more classical,” but the oracle was always still randomized or “in-place” or had restrictions on how it could be queried.

(7) Oxford Ionics (which is now owned by IonQ) announced a 2-qubit gate with 99.99% fidelity: a record, and significantly past the threshold for quantum fault-tolerance. However, as far as I know, it remains to demonstrate this sort of fidelity in a large programmable system with dozens of qubits and hundreds of gates.

(8) Semi-announcement: Quanta reports that “Physicists Take the Imaginary Numbers Out of Quantum Mechanics,” and this seems to have gone viral on my social media. The article misses the opportunity to explain that “taking the imaginary numbers out” is as trivial as choosing to call each complex amplitude “just an ordered pair of reals, obeying such-and-such rules, which happen to mimic the rules for complex numbers.” Thus, the only interesting question here is whether one can take imaginary numbers out of QM in various more-or-less “natural” ways: a technical debate that the recent papers are pushing forward. For what it’s worth, I don’t expect that anything coming out of this line of work will ever be “natural” enough for me to stop explaining QM in terms of complex numbers in my undergraduate class, for example.

(9) The list of accepted talks for the annual QIP conference, to be held January 24-30 in Riga, Latvia, is now out. Lots of great stuff as always.

(10) There are probably other major recent developments in QC that I should’ve put into this post but forgot about. You can remind me about them in the comments.

(11) Indeed there are! I completely forgot that Phasecraft announced two simulations of fermionic systems that might achieve quantum advantage, one using Google’s Willow superconducting chip and the other using a Quantinuum device.


To summarize three takeaways:

  • Evidence continues to pile up that we are not living in the universe of Gil Kalai and the other quantum computing skeptics. Indeed, given the current staggering rate of hardware progress, I now think it’s a live possibility that we’ll have a fault-tolerant quantum computer running Shor’s algorithm before the next US presidential election. And I say that not only because of the possibility of the next US presidential election getting cancelled, or preempted by runaway superintelligence!
  • OK, but what will those quantum computers be useful for? Anyone who’s been reading this blog for the past 20 years, or any non-negligible fraction thereof, hopefully already has a calibrated sense of that, so I won’t belabor. But briefly: yes, our knowledge of useful quantum algorithms has slowly been expanding over the past thirty years. The central difficulty is that our knowledge of useful classical algorithms has also been expanding, and the only thing that matters is the differential between the two! I’d say that the two biggest known application areas for QC remain (a) quantum simulation and (b) the breaking of public-key cryptography, just as they were thirty years ago. In any case, none of the exciting developments that I’ve chosen to highlight in this post directly address the “what is it good for?” question, with the exception of the certified randomness thing.
  • In talks over the past three years, I’ve advocated “verifiable quantum supremacy on current hardware” as perhaps the central challenge right now for quantum computing theory. (As I love to point out, we do know how to achieve any two of (a) quantum supremacy that’s (b) verifiable and (c) runs on current hardware!) So I’m gratified that three of the recent developments that I chose to highlight, namely (1), (4), and (5), directly address this challenge. Of course, we’re not yet sure whether any of these three attempts will stand—that is, whether they’ll resist all attempts to simulate them classically. But the more serious shots on goal we have (and all three of these are quite serious), the better the chances that at least one will stand! So I’m glad that people are sticking their necks out, proposing these things, and honestly communicating what they know and don’t know about them: this is exactly what I’d hoped would happen. Of course, complexity-theoretic analysis of these proposals would also be great, perhaps from people with more youth and/or energy than me. Now it’s time for me to sleep.

Matt von HippelReminder to Physics Popularizers: “Discover” Is a Technical Term

When a word has both an everyday meaning and a technical meaning, it can cause no end of confusion.

I’ve written about this before using one of the most common examples, the word “model”, which means something quite different in the phrases “large language model”, “animal model for Alzheimer’s” and “model train”. And I’ve written about running into this kind of confusion at the beginning of my PhD, with the word “effective”.

But there is one example I see crop up again and again, even with otherwise skilled science communicators. It’s the word “discover”.

“Discover”, in physics, has a technical meaning. It’s a first-ever observation of something, with an associated standard of evidence. In this sense, the LHC discovered the Higgs boson in 2012, and LIGO discovered gravitational waves in 2015. And there are discoveries we can anticipate, like the cosmic neutrino background.

But of course, “discover” has a meaning in everyday English, too.

You probably think I’m going to say that “discover”, in everyday English, doesn’t have the same statistical standards it does in physics. That’s true of course, but it’s also pretty obvious, I don’t think it’s confusing anybody.

Rather, there is a much more important difference that physicists often forget: in everyday English, a discovery is a surprise.

“Discover”, a word arguably popularized by Columbus’s discovery of the Americas, is used pretty much exclusively to refer to learning about something you did not know about yet. It can be minor, like discovering a stick of gum you forgot, or dramatic, like discovering you’ve been transformed into a giant insect.

Now, as a scientist, you might say that everything that hasn’t yet been observed is unknown, ready for discovery. We didn’t know that the Higgs boson existed before the LHC, and we don’t know yet that there is a cosmic neutrino background.

But just because we don’t know something in a technical sense, doesn’t mean it’s surprising. And if something isn’t surprising at all, then in everyday, colloquial English, people don’t call it a discovery. You don’t “discover” that the store has milk today, even if they sometimes run out. You don’t “discover” that a movie is fun, if you went because you heard reviews claim it would be, even if the reviews might have been wrong. You don’t “discover” something you already expect.

At best, maybe you could “discover” something controversial. If you expect to find a lost city of gold, and everyone says you’re crazy, then fine, you can discover the lost city of gold. But if everyone agrees that there is probably a lost city of gold there? Then in everyday English, it would be very strange to say that you were the one who discovered it.

With this in mind, the way physicists use the word “discover” can cause a lot of confusion. It can make people think, as with gravitational waves, that a “discovery” is something totally new, that we weren’t pretty confident before LIGO that gravitational waves exist. And it can make people get jaded, and think physicists are overhyping, talking about “discovering” this or that particle physics fact because an experiment once again did exactly what it was expected to.

My recommendation? If you’re writing for the general public, use other words. The LHC “decisively detected” the Higgs boson. We expect to see “direct evidence” of the cosmic neutrino background. “Discover” has baggage, and should be used with care.

Matt Strassler Event with Professor Daniel Whiteson on Monday November 17 at 7pm

Next Monday, November 17th at 7pm, I’ll be at the Harvard Bookstore with particle physicist and author Daniel Whiteson. Professor Whiteson and his co-author Andy Warner have a nice new book, for the general science-aware reader, exploring an age-old and unanswered question: how universal is the knowledge and understanding that we call “physics”? How much of modern physics is actually telling us about the universe, and how much of it is created by, or an accident of, the humans who have helped bring it about?

For instance, if we started all over again and reran history from scratch, would the physics (and science more generally) of this re-run culture look much like our own, or might it turn out very differently? If another culture on Earth had had time to develop highly mature science (or something like it) in its own direction, independent of Western Europe’s influence, how different might that science be? (Indeed, would our word “science” even be translatable into their worldview?) Or if we encountered aliens with far greater understanding of the universe than we have, would we be able to recognize, parse, grok, appreciate, comprehend, and/or otherwise make sense of their notions of scientific knowledge?

Whiteson and his co-author, wanting to write a popular book rather than a scholarly one, and desiring nevertheless to take on these serious and challenging intellectual questions, have set their focus mostly on the aliens, accompanied by amusing cartoons and a generous helping of dad jokes (hey, some dad jokes are actually very funny.) They’re looking for a broad audience, and hopefully they will get it. But don’t let the light-hearted title (“Do Aliens Speak Physics?“) or the charmingly goofy cover fool you: this book might well make you laugh, but I guarantee it will make you think. Whether you’re just curious about science or you’ve been doing science yourself for years, I suspect that, within the vast array of problems and issues that are raised in this broad-minded book, there will be some you’ve never thought of.

Among scientists and philosophers, there are some who believe that any aliens with the capacity to reach the Earth will obviously “speak physics” — that math and physics float above contingencies of culture and species, and will easily be translated from any intelligent creature to any other. But are they perhaps flying too high? It’s clear that Whiteson and Warner are aiming to poke some holes — lots of holes —- in their hot-air balloon, and to do so in a way that a wide variety of readers can appreciate and enjoy.

I tend to agree with Whiteson on a lot of these issues, but that won’t stop me from asking him some tough questions. You can ask him some tough questions too, if you like — just come to the Harvard Bookstore at 7:00 on Monday and join the conversation!

November 10, 2025

John BaezThe Standard Model – Part 3

Physics is really bizarre and wonderful. Here I start explaining why the Standard Model has U(1) × SU(2) × SU(3) as its symmetry group. But I don’t assume you know anything about groups or quantum mechanics! So I have to start at the beginning: how the electromagnetic, weak, and strong force are connected to the numbers 1, 2, and 3. It’s all about quunits, qubits and qutrits.

You’ve heard of bits, which describe a binary alternative, like 0 and 1. You’ve probably heard about qubits, which are the quantum version of bits. The weak force is connected to qubits where the 2 choices are called “isospin up” and “isospin down”. The most familiar example is the choice between a proton and a neutron. A better example is the choice between an up quark and a down quark.

The strong force is connected to qutrits—the quantum version of a choice between 3 alternatives. In physics these are whimsically called “red”, “green” and “blue”. Quarks come in 3 colors like this.

The electromagnetic force is connected to “quunits” – the quantum version of a choice between just one alternative. It may seem like that’s no choice at all! But quantum mechanics is weird: there’s just one choice, but you can still rotate that choice.

Yes, I know this stuff sounds crazy. But this is how the world actually works. I start explaining it here, and I’ll keep on until it’s all laid out quite precisely.

November 09, 2025

Tommaso DorigoRestoring The Value Of Truth

Truth is under attack. It has always been, of course, because truth has always been a mortal enemy for those who attempt to seize or keep power in their hands. But the amplification of the phenomenon by today's information technology is extremely worrisome. AI today can generate fake videos and images that even experts have trouble flagging as such. This, combined with the different news value and propagation potential of false information with respect to typically less attention-grabbing true facts has created an explosive situation. What to do?

read more

November 08, 2025

Doug NatelsonVortices everywhere

The 2026 APS Oliver E. Buckley Prize in condensed matter physics was announced this week, and it's a really interesting combination of topics that, to a lay person, may seem to be completely unrelated.  

Fig. 1 from this follow-up PRB.
On the one hand, John Reppy (at age 94!) and Dave Bishop were honored for their work examining the properties of vortices in thin films of superfluid helium-4.  Relevant papers include this one from 1977, where they used a torsion pendulum coated with the helium film to examine the transition between normal and superfluid.  When the helium becomes a superfluid, it has (at low speeds) no viscosity, so it no longer has to rotate with the torsion pendulum; this means the rotational moment of inertia goes from that of (pendulum+helium) to just (pendulum), and the period of the oscillations increases.  Really detailed measurements of the oscillations and their damping allowed Reppy and Bishop to compare with models of the superfluid transition based on work by Kosterlitz and Thouless (and Berezinskii).  See the image for a diagram of the experimental setup - very clever and intricate.  

The key idea here is the role of vortices.  Superfluidity in helium is described by an order parameter that looks like a wavefunction - it has an amplitude, \(\Psi_{0}\), and a phase \(\phi\), so that \(\Psi(\mathbf{r}) = \Psi_{0} \exp(i \phi)\).   That order parameter is supposed to be single-valued, meaning if you go around a closed loop of some kind, that phase will either remain the same or ramp by some integer multiple of \(2\pi\).  The gradient of the phase is related to the velocity of the superfluid, so if the phase winds by \(2\pi\), that implies there is a circulation of flow and orbital angular momentum that has to be an integer multiple of \(\hbar\).  In the BKT theory, the demise of the superfluid phase as the system is warmed happens through the creation and unbinding of vortex-antivortex pairs.

On the other hand, the other recipients of the Buckley Prize were Gwendal Fève and Mike Manfra for their work (experiments here and here) regarding the braiding statistics of anyons in fractional quantum Hall systems.  I'd written about anyons here.  For electrons in 2D, the wavefunctions of excitations of the fractional quantum Hall system look like vortices.  The phase of the electronic wavefunction can wind due to circulation, and because electrons are charged, the phase can also wind due to magnetic flux attached to the little whirlpool.  It's the combination of these phase effects that can lead to those excitations acting like anyons (so that when two are physically swapped or braided around one another, the wavefunction picks up a phase factor that is not just the \(+1\) of bosons or the \(-1\) of fermions).  

As my friend Dan Arovas pointed out, there was a hope back in the early 1980s that perhaps vortices in superfluid helium would also act like anyons and have fractional statistics.  However, this paper by Haldane and Wu disproved that possibility.  

Vortex shedding, from here.
Because of the relationship between quantum phase winding and actual flow of (density) currents, vortices show up in lots of places in hard condensed matter physics.  Classical vortices are also physically nontrivial objects - they're topological and often seem to have very counterintuitive properties and motions.  Heck, Lord Kelvin was so taken by this that he thought (pre-quantum) that maybe everything is really vortices of some kind.  

Perhaps it is fitting that I am posting this on the 85th anniversary of the Tacoma Narrows bridge collapse.  That classic civil engineering failure was caused by vortex shedding by the bridge coupling to its torsional resonance frequency.  Vortices can have big consequences!  


November 07, 2025

Matt von HippelExplain/Teach/Advocate

Scientists have different goals when they communicate, leading to different styles, or registers, of communication. If you don’t notice what register a scientist is using, you might think they’re saying something they’re not. And if you notice someone using the wrong register for a situation, they may not actually be a scientist.

Sometimes, a scientist is trying to explain an idea to the general public. The point of these explanations is to give you appreciation and intuition for the science, not to understand it in detail. This register makes heavy use of metaphors, and sometimes also slogans. It should almost never be taken literally, and a contradiction between two different scientist explanations usually just means they are using incompatible metaphors for the same concept. Sometimes, scientists who do this a lot will comment on other metaphors you might have heard, referencing other slogans to help explain what those explanations miss. They do this knowing that they do, in the end, agree on the actual science: they’re just trying to give you another metaphor, with a deeper intuition for a neglected part of the story.

Other times, scientists are trying to teach a student to be able to do something. Teaching can use metaphors or slogans as introductions, but quickly moves past them, because it wants to show the students something they can use: an equation, a diagram, a classification. If a scientist shows you any of these equations/diagrams/classifications without explaining what they mean, then you’re not the student they had in mind: they had designed their lesson for someone who already knew those things. Teaching may convey the kinds of appreciation and intuition that explanations for the general public do, but that goal gets much less emphasis. The main goal is for students with the appropriate background to learn to do something new.

Finally, sometimes scientists are trying to advocate for a scientific point. In this register, and only in this register, are they trying to convince people who don’t already trust them. This kind of communication can include metaphors and slogans as decoration, but the bulk will be filled with details, and those details should constitute evidence: they should be a structured argument, one that lays out, scientifically, why others should come to the same conclusion.

A piece that tries to address multiple audiences can move between registers in a clean way. But if the register jumps back and forth, or if the wrong register is being used for a task, that usually means trouble. That trouble can be simple boredom, like a scientist’s typical conference talk that can’t decide whether it just wants other scientists to appreciate the work, whether it wants to teach them enough to actually use it, or whether it needs to convince any skeptics. It can also be more sinister: a lot of crackpots write pieces that are ostensibly aimed at convincing other scientists, but are almost entirely metaphors and slogans, pieces good at tugging on the general public’s intuition without actually giving scientists anything meaningful to engage with.

If you’re writing, or speaking, know what register you need to use to do what you’re trying to do! And if you run into a piece that doesn’t make sense, consider that it might be in a different register than you thought.

November 06, 2025

n-Category Café The Inverse Cube Force Law

Here’s a draft of my next column for the Notices of the American Mathematical Society. It’s about the inverse cube force law in classical mechanics.

Newton’s Principia is famous for his investigations of the inverse square force law for gravity. But in this book Newton also did something that was rarely discussed until the 1990s. He figured out what kind of central force exerted upon a particle can rescale its angular velocity by a constant factor without affecting its radial motion. This turns out to be a force obeying an inverse cube law.

Given a particle in Euclidean space, a central force is a force that points toward or away from the origin and depends only on the particle’s distance from the origin. If the particle’s position at time is r(t) n\mathbf{r}(t) \in \mathbb{R}^n and its mass is some number m>0,m \gt 0, we have

mr¨(t)=F(r(t))r^(t) m \, \ddot{\mathbf{r}}(t) = F(r(t)) \,\hat{\mathbf{r}}(t)

where r^(t)\hat{\mathbf{r}}(t) is a unit vector pointing outward from the origin at the point r(t).\mathbf{r}(t). A particle obeying this equation always moves in a plane through the origin, so we can use polar coordinates and write the particle’s position as (r(t),θ(t).(r(t), \theta(t). With some calculation one can show the particle’s distance from the origin, r(t),r(t), obeys

mr¨(t)=F(r(t))+L 2mr(t) 3(1) m \ddot r(t) = F(r(t)) + \frac{L^2}{m r(t)^3 } \qquad \qquad (1)

Here L=mr(t) 2θ˙(t)L = m r(t)^2 \dot \theta(t), the particle’s angular momentum, is constant in time. The second term in the equation above says that the particle’s distance from the origin changes as if there were an additional force pushing it outward. This is a “fictitious force”, an artifact of working in polar coordinates. It is called the centrifugal force. And it obeys an inverse cube force law!

This explains Newton’s observation. Let us see why. Suppose we have two particles moving in two different central forces F 1F_1 and F 2,F_2, each obeying a version of equation (1), with the same mass mm and the same radial motion r(t),r(t), but different angular momenta L 1L_1 and L 2.L_2. Then we must have

F 1(r(t))+L 1 2mr(t) 3=F 2(r(t))+L 2 2mr(t) 3 F_1(r(t)) + \frac{L_1^2}{m r (t)^3} = F_2(r(t)) + \frac{L_2^2}{m r(t)^3}

If the particle’s angular velocities are proportional we must have L 2=kL 1L_2 = k L_1 for some constant k,k, so

F 2(r 1(t))F 1(r(t))=(k1)L 1mr(t) 3 F_2(r_1(t)) - F_1(r(t)) = \frac{(k - 1)L_1}{m r (t)^3}

This says that F 2F_2 equals F 1F_1 plus an additional inverse cube force.

There are other interesting things about the inverse cube force law. Newtonian gravity is an attractive inverse square force, say F(r)=c/r 2F(r) = -c/r^2 with c>0,c \gt 0, so in this case we have

mr¨(t)=c/r(t) 2+L 2mr(t) 3 m \ddot r(t) = -c/r(t)^2 + \frac{L^2}{m r(t)^3 }

Because 1/r 31/r^3 grows faster than 1/r 21/r^2 as r0,r \downarrow 0, as long as the angular momentum LL is nonzero the repulsion of the centrifugal force will beat the attraction of gravity for sufficiently small r,r, and the particle will not fall in to the origin. The same is true for any attractive force F(r)=c/r pF(r) = -c/r^p with p<3.p \lt 3. But an attractive inverse cube force can overcome the centrifugal force and make a particle fall in to the origin.

In fact there are three qualitatively different possibilities for the motion of a particle in an attractive inverse cube force F(r)=c/r 3,F(r) = -c/r^3, depending on the value of cc. With work we can solve for 1/r1/r as a function of θ\theta (which is easier than solving for rr). There are three cases depending on the value of

ω 2=1cmL 2 \omega^2 = 1 - \frac{c m}{L^2}

They are vaguely analogous to the elliptical, parabolic and hyperbolic orbits of a particle in an inverse square force law:

1r(θ)={Acos(ωθ)+Bsin(ωθ) if ω 2>0 A+Bθ if ω=0 Ae ωθ+Be ωθ if ω 2<0 \frac{1}{r(\theta)} = \left\{ \begin{array}{lcl} A \cos(\omega \theta) + B \sin(\omega \theta) & \text{if} & \omega^2 \gt 0 \\ \\ A + B \theta & \text{if} & \omega = 0 \\ \\ A e^{\omega \theta} + B e^{-\omega \theta} & \text{if} & \omega^2 \lt 0 \end{array} \right.

The third case occurs when the attractive inverse cube force is strong enough to overcome the centrifugal force: c>L 2/m.c \gt L^2/m. Then the particle can spiral in to its doom, hitting the origin in a finite amount of time after infinitely many orbits, like this:

All three curves in the equation above are called Cotes spirals, after Roger Cotes’ work on the inverse cube force law, published posthumously in 1722. Cotes seems to have been the first to compute the derivative of the sine function. After Cotes’ death at the age of 33, Newton supposedly said “If he had lived we would have known something.”

The subtlety of the inverse cube force law is vastly heightened when we study it using quantum rather than classical mechanics. Here if cc is too large the theory is ill-defined, because there is no reasonable choice of self-adjoint Hamiltonian. If cc is smaller the theory is well-behaved. But at a certain borderline point it exhibits a remarkable property: spontaneous breaking of scaling symmetry. I hope to discuss this in my next column.

For more on the inverse cube force law, see:

  • N. Grossman, The Sheer Joy of Celestial Mechanics, Birkhäuser, Basel, 1996, p. 34.

For more on Newton’s work involving the inverse cube force law, see:

Cotes’ book is

  • Roger Cotes, Harmonia Mensuarum, Cambridge, 1722.

November 04, 2025

n-Category Café Dynamics in Jordan Algebras

In ordinary quantum mechanics, in the special case where observables are described as self-adjoint n×nn \times n complex matrices, we can describe time evolution of an observable O(t)O(t) using Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = -i [H, O(t)]

where HH is a fixed self-adjoint matrix called the Hamiltonian. This framework is great when we want to focus on observables rather than states. But Heisenberg’s equation doesn’t make sense in a general Jordan algebra. In this stripped-down framework, all we can do is raise observables to powers and take real linear combinations of them. This lets us define a ‘Jordan product’ of observables:

AB=12((A+B) 2A 2B 2)=12(AB+BA) A \circ B = \frac{1}{2} ((A + B)^2 - A^2 - B^2) = \frac{1}{2} (A B + B A)

but not commutators and not multiplication by ii. What do we do then?

I wrote a long paper about this:

My starting-point was that self-adjoint complex matrices form not only a Jordan algebra with product

AB=12(AB+BA) A \circ B = \frac{1}{2} (A B + B A)

but also a Lie algebra with bracket

i[A,B]=i(ABBA) -i [A, B] = -i(A B - B A)

See, the commutator of two self-adjoint matrices is skew-adjoint, but we can multiply it by ii, or more conventionally i-i, to get something self-adjoint. That’s what is going on in Heisenberg’s equation. But this trick doesn’t work for other Jordan algebras, at least not automatically—so there was a lot to say.

I just bumped into a nice paper on this issue that I hadn’t seen before:

The idea here is pretty wild: you can replace the commutator in Heisenberg’s equation by an associator:

(A,B,C)=(AB)CA(BC) (A, B, C) = (A \circ B) \circ C - A \circ (B \circ C)

This is well-defined whenever our observables are elements in a Jordan algebra. Jordan algebras are always commutative, but rarely associative!

Here’s the trick. Let 𝔥 n()\mathfrak{h}_n(\mathbb{C}) be the Jordan algebra of self-adjoint n×nn \times n complex matrices, and let’s start with Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = -i [H, O(t)]

where H𝔥 n()H \in \mathfrak{h}_n(\mathbb{C}). Suppose we can write

H=4i[X,Y] H = -4i [X, Y]

for some X,Y𝔥 n()X, Y \in \mathfrak{h}_n(\mathbb{C}). In this case we can use a really cool identity to express the commutator in Heisenberg’s equation in terms of an associator:

[[X,Y],A]=14(X,A,Y) [[X, Y], A] = -\frac{1}{4}(X, A, Y)

This holds in any associative algebra if you define [X,Y]=XYYX[X,Y] = X Y - Y X, XY=12(XY+YX)X \circ Y = \tfrac{1}{2} (X Y + Y X) and (X,A,Y)=(X, A, Y) = (XA)YX(AY) (X \circ A) \circ Y - X \circ (A \circ Y). It’s easy to check: just expand out both sides and compare them!

Using this identity, we get

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

Now we’re describing dynamics using only operations that are available in any Jordan algebra!

This raises the question of when a self-adjoint complex matrix HH can be written as 4i[X,Y]-4i [X, Y] for self-adjoint matrices X,YX, Y. This is true whenever HH is traceless, since 𝔰𝔲(n)\mathfrak{su}(n) is a compact simple real Lie algebra, and every element of such a Lie algebra is a commutator (as shown by Akhieser).

But any self-adjoint complex matrix HH is of the form H+λIH' + \lambda I where HH' is traceless, so writing H=4i[X,Y]H' = -4i[X,Y] we have

[H,O(t)]=[H+λI,O(t)]=[H,O(t)] [H, O(t)] = [H' + \lambda I, O(t)] = [H', O(t)] =4i[[X,Y],O(t)]=i(X,O(t),Y) = -4i [[X,Y], O(t)] = i (X, O(t), Y)

so we can rewrite Heisenberg’s equation as

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

Moreover, in any Jordan algebra, any pair of elements X,YX, Y determines a derivation (X,,Y)(X, \cdot, Y): see Section I.7 of Jacobson’s Structure and Representations of Jordan Algebras. In the finite-dimensional case there is no difficulty with exponentiating any derivation to obtain a one-parameter group of automorphisms. Thus, for any elements X,YX, Y of a finite-dimensional Jordan algebra, the solution of the above equation always determines a one-parameter group of Jordan algebra automorphisms! And this is just what we’d want for describing how observables change with time.

The are two obvious next questions: one mathematical and one more philosophical.

First, how many one-parameter groups of Jordan algebra automorphisms do we actually get out of solutions to

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

In the case of 𝔥 n()\mathfrak{h}_n(\mathbb{C}), we get them all, since it’s already known that we get them all from Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = - i [H , O(t)]

What about 𝔥 n()\mathfrak{h}_n(\mathbb{R}) and 𝔥 n()\mathfrak{h}_n(\mathbb{H})? I’m actually more interested in the exceptional Jordan algebra 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}), and here it seems we get them all! This was shown in a paper that’s fairly hard to find even though it’s available for free online:

It starts on page 214 of the PDF file.

(The editor of this journal has some crazy ideas, which has put off some people I’m talking to about this paper. But you can’t judge a paper by the journal it appeared in. Truini and Biedenharn are good — in fact Biedenharn is famous for helping discover an identity, the Biedenharn–Elliott identity, that amounts to the pentagon identity for the category of representations of SU(2)\text{SU}(2)! And the paper looks fine, as far as I can tell.)

Second, the more philosophical question: what does it mean to describe dynamics using not one observable, the Hamiltonian, but two? Perhaps the best way to tackle this is to try doing it, and seeing how it works. Note that this method is not just good for dynamics, but for any Lie group of symmetries.

November 02, 2025

Doug NatelsonInteresting preprints: chirality-induced spin selectivity + quantum gravity

This continues to be a very busy time, but I wanted to point out two preprints that caught my eye this week.  Their subjects are completely disparate, but they stand out as essentially reviews written in a much more conversational tone than the usual literature.

The first is this preprint about chirality-induced spin selectivity, a subject that I've mentioned before on this blog.  There is now an extensive body of evidence (of varying quality) that there is a connection between structural chirality of molecules and their interactions with the spin angular momentum of electrons.  This includes monolayers of chiral molecules leading to net spin polarization of photoemitted electrons (here), a lot of electronic transport experiments involving chiral molecules and magnetic electrodes that seem to show spin-dependent transmission that is absent with achiral molecules, and even a chirality dependence of molecular adsorption kinetics on magnetic surfaces (here).  The preprint is a provocative discussion of the topic and possible mechanisms, and the importance of precision in the description of the various phenomena.

On a completely different topic, this preprint is a fun discussion about quantum gravity (!) and how condensed matter ideas of "the vacuum" can lead to insights about how quantum mechanics and gravity might need to play together.  One fun bit early on is a discussion of something I like to point out to my undergrad stat mech students:  A single hydrogen atom in a very very large box will apparently (if the usual stat mech formalism of partition functions is valid) be spontaneously ionized, even when the box (which presumably functions as a reservoir at temperature \(T\)) and atom are at temperatures faaaaaar below the energy scale for ionization.  This is discussed nicely in this 1966 article in the Journal of Chemical Education.  Anyway, I thought this was an interesting discussion from three condensed matter theorists.

November 01, 2025

Tommaso DorigoNovember First

Today is November 1st, the day dedicated to the dead, and I am in northern Sweden where daylight is scarce this time of the year. The two things conjure to arise thoughts of a darkish nature. 


[Above, a lousy picture taken this evening in a cemetery in Gammelstad, close to Lulea, in Norrbotten, Sweden. Sorry for the bad quality... Yet the landscape with all those small lights was really inspiring.]

read more

October 27, 2025

John PreskillThe sequel

This October, fantasy readers are devouring a sequel: the final installment in Philip Pullman’s trilogy The Book of Dust. The series follows student Lyra Silvertongue as she journeys from Oxford to the far east. Her story features alternate worlds, souls that materialize as talking animals, and a whiff of steampunk. We first met Lyra in the His Dark Materials trilogy, which Pullman began publishing in 1995. So some readers have been awaiting the final Book of Dust volume for 30 years. 

Another sequel debuts this fall. It won’t spur tens of thousands of sales; nor will Michael Sheen narrate an audiobook version of it. Nevertheless, the sequel should provoke as much thought as Pullman’s: the sequel to the Maryland Quantum-Thermodynamics Hub’s first three years.

More deserving of a Carnegie Medal than our hub, but the hub deserves no less enthusiasm!

The Maryland Quantum-Thermodynamics Hub debuted in 2022, courtesy of a grant from the John F. Templeton Foundation. Six theorists, three based in Maryland, have formed the hub’s core. Our mission has included three prongs: research, community building, and outreach. During the preceding decade, quantum thermodynamics had exploded, but mostly outside North America. We aimed to provide a lodestone for the continent’s quantum-thermodynamics researchers and visitors.

Also, we aimed to identify the thermodynamics of how everyday, classical physics emerges from quantum physics. Quantum physics is reversible (doesn’t distinguish the past from the future), is delicate (measuring a quantum system can disturb it), and features counterintuitive phenomena such as entanglement. In contrast, our everyday experiences include irreversibility (time has an arrow), objectivity (if you and I read this article, we should agree about its contents), and no entanglement. How does quantum physics give rise to classical physics at large energy and length scales? Thermodynamics has traditionally described macroscopic, emergent properties. So quantum thermodynamics should inform our understanding of classical reality’s emergence from quantum mechanics.

Our team has approached this opportunity from three perspectives. One perspective centers on quantum Darwinism, a framework for quantifying how interactions spread information about an observed quantum system. Another perspective highlights decoherence, the contamination of a quantum system by its environment. The third perspective features incompatible exchanged quantities, described in an earlier blog post. Or two. Or at least seven

Each perspective led us to discover a tension, or apparent contradiction, that needs resolving. One might complain that we failed to clinch a quantum-thermodynamic theory of the emergence of classical reality. But physicists adore apparent contradictions as publishers love splashing “New York Times bestseller” on their book covers. So we aim to resolve the tensions over the next three years.

Physicists savor paradoxes and their ilk.

I’ll illustrate the tensions with incompatible exchanged quantities, of course. Physicists often imagine a small system, such as a quantum computer, interacting with a large environment, such as the surrounding air and the table on which the quantum computer sits. The system and environment may exchange energy, particles, electric charge, etc. Typically, the small system thermalizes, or reaches a state mostly independent of its initial conditions. For example, after exchanging enough energy with its environment, the system ends up at the environment’s temperature, mostly regardless of the system’s initial temperature. 

For decades, physicists implicitly assumed that the exchanged quantities are compatible: one can measure them simultaneously. But one can’t measure all of a quantum system’s properties simultaneously. Position and momentum form the most famous examples. Incompatibility epitomizes quantum physics, underlying Heisenberg’s uncertainty relation, quantum error correction, and more. So collaborators and I ask how exchanged quantities’ incompatibility alters thermalization, which helps account for time’s arrow. 

Our community has discovered that such incompatibility can hinder certain facets of thermalization—in a sense, stave off certain aspects of certain quantum systems’ experience of time. But incompatible exchanged quantities enhance other features of thermalization. How shall we reconcile the hindrances with the enhancements? Does one of the two effects win out? I hope to report back in three years. For now, I’m rooting for Team Hindrance.

In addition to resolving apparent conflicts, we’re adding a fourth perspective to our quiver—a gravitational one. In our everyday experiences, space-time appears smooth; unlike Lyra’s companion Will in The Subtle Knife, we don’t find windows onto other worlds. But quantum physics, combined with general relativity, suggests that you’d find spikes and dips upon probing space-time over extremely short length scales. How does smooth space-time emerge from its quantum underpinnings? Again, quantum thermodynamics should help us understand.

To address these challenges, we’re expanding the hub’s cast of characters. The initial cast included six theorists. Two more are joining the crew, together with the hub’s first two experimentalists. So is our first creative-writing instructor, who works at the University of Maryland (UMD) Jiménez-Porter Writers’ House.

As the hub has grown, so has the continent’s quantum-thermodynamics community. We aim to continue expanding that community and strengthening its ties to counterparts abroad. As Lyra learned in Pullman’s previous novel, partnering with Welsh miners and Czech book sellers and Smyrnan princesses can further one’s quest. I don’t expect the Maryland Quantum-Thermodynamics Hub to attract Smyrnan princesses, but a girl can dream. The hub is already partnering with the John F. Templeton Foundation, Normal Computing, the Fidelity Center for Applied Technology, the National Quantum Laboratory, Maryland’s Capital of Quantum team, and more. We aim to integrate quantum thermodynamics into North America’s scientific infrastructure, so that the field thrives here even after our new grant terminates. Reach out if you’d like to partner with us.

To unite our community, the hub will host a gathering—a symposium or conference—each year. One conference will feature quantum thermodynamics and quantum-steampunk creative writing. Scientists and authors will present. We hope that both groups will inspire each other, as physicist David Deutsch’s work on the many-worlds formulation of quantum theory inspired Pullman.

That conference will follow a quantum-steampunk creative-writing course to take place at UMD during spring 2026. I’ll co-teach the course with creative-writing instructor Edward Daschle. Students will study quantum thermodynamics, read published science-fiction stories, write quantum-steampunk stories, and critique each other’s writing. Five departments have cross-listed the course: physics, arts and humanities, computer science, chemistry, and mechanical engineering. If you’re a UMD student, you can sign up in a few weeks. Do so early; seats are limited! We welcome graduate students and undergrads, the latter of whom can earn a GSSP general-education credit.1 Through the course, the hub will spread quantum thermodynamics into Pullman’s world—into literature.

Pullman has entitled his latest novel The Rose Field. The final word refers to an object studied by physicists. A field, such as an electric or gravitational field, is a physical influence spread across space. Hence fiction is mirroring physics—and physics can take its cue from literature. As ardently as Lyra pursues the mysterious particle called Dust, the Maryland Quantum-Thermodynamics Hub is pursuing a thermodynamic understanding of the classical world’s emergence from quantum physics. And I think our mission sounds as enthralling as Lyra’s. So keep an eye on the hub for physics, community activities, and stories. The telling of Lyra’s tale may end this month, but the telling of the hub’s doesn’t.

1Just don’t ask me what GSSP stands for.

October 15, 2025

Clifford JohnsonNobel Prize in Physics 2025: Who/What/Why

I started a tradition a little while back where every year we have a special departmental colloquium entitled "The Nobel Prize in Physics: Who/What/Why". This year my job in finding speakers was made easier by having 2/3 of this years newly-minted Nobel Prize winners in physics in the Department! (Michel Devoret and John Martinis.) So our room was a bit more well-attended than normal...(hundreds and hundreds rather than dozens and dozens). Here is a recording of the event, which I was delighted to host, and there's a celebration afterwards too. (Pls share widely!)
[...] Click to continue reading this post

The post Nobel Prize in Physics 2025: Who/What/Why appeared first on Asymptotia.

October 01, 2025

Robert HellingHolosplit

 Recently I had to update Mathematica on my laptop and after having solved the challenges of the license manager that keeps looking different every time I have to use it, I learned that Mathematica 14 can now officially work with finite fields.

This reminded me that for a while I wanted to revive an old project that had vanished together with the hard drive of some old computer: Holosplit. So, over the last two days and with the help of said version of Mathematica I did a complete rewrite which you can now find on Github.

It consists of two C programs "holosplit" and "holojoin". To the first you give a positive integer \(N\) and a file and it spits out a new file ("fragment") that is roughly \(1/N\) of the size. Every time you do that you obtain a new random fragment.

The later you give any collection of \(N\) of these fragments and it reproduces the original file. So you can for example distribute a file over 10 people such that when any 3 of them work together, they can recover the original. 

How does it work? I uses the finite field \(F\) of \(2^3=256\) elements (in the Github repository, there is also a header file that implements arithmetic in \(F\) and matrix operations like product and inverse over it). Each time, it is invoked, it picks a random vector \(v\in F^N\) and writes it to the output. Then it reads \(N\) bytes from the file at a time which it also interprets as a vector \(d\in F^N\). It then outputs the byte that corresponds to the scalar product \(v\cdot d\).

To reassemble the file, holojoin takes the \(N\) files with its random vectors \(v_1,\ldots,v_N\) and interprets those as the rows of a \(N\times N\) matrix \(A\). With probability

$$\frac{\prod_{k=1}^N \left(256^N-k\right)}{(256)^{N^2}}$$

which exponentially in \(N\) approaches 1 this matrix is invertible (homework: why?). So we can read one byte from each file, assemble those into yet another vector \(e\in F^N\) and recover

$$d=A^{-1}e.$$

Besides the mathematics, it also poses philosophical/legal questions: Consider for example the original file is copyrighted, for example an mp3 or a video. The fragments are clearly derived works. But individually, they do not contain the original work, without sufficiently many other fragments they are useless (although not in a cryptographic sense). So by publishing one fragment, I do not provide access to the original work. What if others publish other fragments? Then my fragment could be the last remaining one that was missing. If there are more, any individual fragment is redundant so publishing it strictly speaking does not provide new information. 

September 26, 2025

Peter Rohde Photo albums

Peter’s photos: https://www.icloud.com/sharedalbum/#B275oqs3qKSZvQ

Screenshots: https://www.icloud.com/sharedalbum/#B27532ODWjIQb9

Climbing book launch: https://www.icloud.com/sharedalbum/#B27GWZuqDGnuOyN

Salisbury waters: https://www.icloud.com/sharedalbum/#B275qXGF1JQFkx

Christmas with Ash: https://www.icloud.com/sharedalbum/#B27G6XBubAhoT6

Hosin BBQ duck: https://www.icloud.com/sharedalbum/#B27GY8gBYG3b5mD

Hawks Nest to Smiths Lake: https://www.icloud.com/sharedalbum/#B2759UlCqSH5bE

Europe & Alps: https://www.icloud.com/sharedalbum/#B275ON9t3W0lu

Point Perpendicular: https://www.icloud.com/sharedalbum/#B27GqkRUiGivXD2

Newnes canyoning: https://www.icloud.com/sharedalbum/#B27GfnH8tgHSmX

Coffs Harbour to Yamba: https://www.icloud.com/sharedalbum/#B27J0DiRHJKuuWr

Wendy Bruere Christmas (2020): https://www.icloud.com/sharedalbum/#B27G4TcsmGoHysj

Six Foot Track: https://www.icloud.com/sharedalbum/#B2753qWtHZA9EX

Kosciusko to Kiandra: https://www.icloud.com/sharedalbum/#B27GgZLKuGaewVm

Camping food: https://www.icloud.com/sharedalbum/#B27GtnIORgbmHu

The Aardvark: https://www.icloud.com/sharedalbum/#B275VaUrzvmAiT

Kangaroo Valley kayaking: https://www.icloud.com/sharedalbum/#B27JEsNWnJrCpi0

Claustral canyon: https://www.icloud.com/sharedalbum/#B2755Z2WMOTpsk

Budawang: https://www.icloud.com/sharedalbum/#B27GDdyTvGvpINL

Mother’s Day panoramas (2021): https://www.icloud.com/sharedalbum/#B27GFssfGG9WmJP

Point Perpendicular & Nowra: https://www.icloud.com/sharedalbum/#B27GRMtznGPdeuZ

Blood moon: https://www.icloud.com/sharedalbum/#B27GdIshaG8NgGX

La Perouse to Coogee: https://www.icloud.com/sharedalbum/#B275aVbMK4h7qo

Canberra ASPI launch: https://www.icloud.com/sharedalbum/#B27GQOeMmGj4Zcv

Edible foraging: https://www.icloud.com/sharedalbum/#B275ejO179Si0N

Sydney to Wollongong: https://www.icloud.com/sharedalbum/#B275M7GFPUasMe

Album for Dad, Father’s Day (2021): https://www.icloud.com/sharedalbum/#B2752plgjnnkUe

Vaucluse (with Cheryl, Nestor & Wendy): https://www.icloud.com/sharedalbum/#B275CmvAS4uA0Z

Bouddi National Park: https://www.icloud.com/sharedalbum/#B27GdPblXG8WdOo

Tom Thumb (the 2nd): https://www.icloud.com/sharedalbum/#B275aDWbr4CN2w

Eden to Victoria: https://www.icloud.com/sharedalbum/#B27GJDfWGArX8l

Wendy’s book launch (the 2nd): https://www.icloud.com/sharedalbum/#B27GIcgc2G7h08y

Mark & Pat Bruere visit Sydney: https://www.icloud.com/sharedalbum/#B27G0ehgLbyWyg

New Years Eve climb (2021): https://www.icloud.com/sharedalbum/#B27Ju8EH6JOZxmU

Newnes Canyoning (2022): https://www.icloud.com/sharedalbum/#B275BydzFU0GZ8

Royal National Park (2022): https://www.icloud.com/sharedalbum/#B27GlxzuqGVI5nE

Peter & Wendy: https://www.icloud.com/sharedalbum/#B27Gf693ZG52tfd

Book photo shoots: too rude…

Wendy & Peter’s mushroom trip: https://www.icloud.com/sharedalbum/#B27GrhkPxG27So8

Post-mushroom hike: https://www.icloud.com/sharedalbum/#B27GdFryYG8i3Ur

Wendy Kalymnos favourites: https://www.icloud.com/sharedalbum/#B27JqstnBJEXkH2

Wendy Frenchmans screenshots: https://www.icloud.com/sharedalbum/#B27Jr1PPdJpd7Dq

Instagram: https://www.icloud.com/sharedalbum/#B27GzFCC1Gb4tqr

Haute route: https://www.icloud.com/sharedalbum/#B27J8GySPJtWoQ1

Kim’s KKKalendar: https://www.icloud.com/sharedalbum/#B275fk75vIL0sH

Frenchmans Cap Wild: https://www.icloud.com/sharedalbum/#B27G4VTwGGoFBkz

Photoshoot with Zixin: https://www.icloud.com/sharedalbum/#B27GPCdxkGKPkM4

Wendy birthday hike (2023): https://www.icloud.com/sharedalbum/#B27GWBC59GnHpQW

Bateman’s Bay to Bawley Point: https://www.icloud.com/sharedalbum/#B27JsHvHoJ8bxWf

Stockton Sand dunes (2023): https://www.icloud.com/sharedalbum/#B27GVfZ2vGloFZV

Wendy book launch (2023): https://www.icloud.com/sharedalbum/#B27J058xyJR4IBM

Dolomites (2023): https://www.icloud.com/sharedalbum/#B0Z5kuVsbGJUzKO

Mount Arapiles: https://www.icloud.com/sharedalbum/#B275GH8Mq8Uh2X

Mount Solitary loop: https://www.icloud.com/sharedalbum/#B275nhQST2mETE

Klaus Hanz Franz Rohde Kunst: https://www.icloud.com/sharedalbum/#B27GqQrCLGiY3vb

Klaus Rohde funeral slideshow: https://www.icloud.com/sharedalbum/#B27GDZLe8GXP58K

Dad (old, B&W): https://www.icloud.com/sharedalbum/#B27GLLXGLJ5mbT2

Klaus & Ursula wedding: https://www.icloud.com/sharedalbum/#B275cLqfN7154g

Test Greece: https://www.icloud.com/sharedalbum/#B27Jq4WnLJ6JMNd

From Will Skea (Alps): https://www.icloud.com/sharedalbum/#B27JHciePJFwacG

From Will Skea (Frenchmans Cap): https://www.icloud.com/sharedalbum/#B275ZhN2v3EVq6

From Will Skea (Arapiles): https://www.icloud.com/sharedalbum/#B27JPrgBGJu3BTD

Coffs Harbour to Yamba (2): https://www.icloud.com/sharedalbum/#B27GFqhgJG9LHgT

Mark magic show (2021): https://www.icloud.com/sharedalbum/#B27G60dj6ARCvd

Wendy Christmas present (2020): https://www.icloud.com/sharedalbum/#B275FrPQ6GxvRu

AHS 25 year reunion: https://www.icloud.com/sharedalbum/#B275O3DjHUvSv

WhatsApp: https://www.icloud.com/sharedalbum/#B275tzEA5fX1nc

Armidale High School: https://www.icloud.com/sharedalbum/#B27GnbeumG4PnAF

Book photos for Mum & Dad: https://www.icloud.com/sharedalbum/#B27Gtec4XQkASe

Miscellaneous: https://www.icloud.com/sharedalbum/#B27Gq6kMgGKn7GR

Three Capes Trail (2022): https://www.icloud.com/sharedalbum/#B27G7HOIlGrDUGZ

Childhood computer programming: https://www.icloud.com/sharedalbum/#B275fu2MutDU8N

Magic with Mark in Maroubra: https://www.icloud.com/sharedalbum/#B27Gv6DhEGD9U3G

Photoshoot with Zixin (2024): https://www.icloud.com/sharedalbum/#B27GCATCnJGoRfW

Butt Crack (2021): https://www.icloud.com/sharedalbum/#B275VtHQfMv0zw

Greece photos new (edited to remove photos from wrong album): https://www.icloud.com/sharedalbum/#B27GY3uThGoBcGj

Singapore (all combined): https://www.icloud.com/sharedalbum/#B275qsTcwJKJjl

Hong Kong (transit): https://www.icloud.com/sharedalbum/#B2759v1AbS8Hve

Taiwan: https://www.icloud.com/sharedalbum/#B27GQD2D7Gw0hAp

India (combined): https://www.icloud.com/sharedalbum/#B27Gtue8VQy83g

Freycinet: https://www.icloud.com/sharedalbum/#B27G5VpecGE5Tbg

Triglav: https://www.icloud.com/sharedalbum/#B275MbK9Vy8erz

Shared with me: https://www.icloud.com/sharedalbum/#B27GGXqixzPOrm

Mount Wellington climbing: https://www.icloud.com/sharedalbum/#B27Gd59qiG8Kjy4

New Zealand combined (2004): https://www.icloud.com/sharedalbum/#B27GIZ8BIGNN5jy

New Zealand combined (2005): https://www.icloud.com/sharedalbum/#B27GcuRfIGFVIcL

Yea: https://www.icloud.com/sharedalbum/#B27GZYbYHGhFIir

Mount Pleasant: https://www.icloud.com/sharedalbum/#B275Iy2hC0JTTL

D’Aguilar: https://www.icloud.com/sharedalbum/#B27Gh7fzTGZBosS

Bali (2001): https://www.icloud.com/sharedalbum/#B27G1qNHBGOTbIr

Samba Ninjas: https://www.icloud.com/sharedalbum/#B27GG34bAzqQ0v

Armidale (misc): https://www.icloud.com/sharedalbum/#B27GSkLVwGyobbX

Emma’s party (2008): https://www.icloud.com/sharedalbum/#B275S2ms99Zyby

Goettingen (2011): https://www.icloud.com/sharedalbum/#B27JIrbT3Jsgxhd

South Coast track: https://www.icloud.com/sharedalbum/#B27G58NWBG6QyN7

Minsk (2006): https://www.icloud.com/sharedalbum/#B27G3JpSBGX1UkQ

Baden-Baden (2019): https://www.icloud.com/sharedalbum/#B27595X5HTVzJr

Berlin (combined): https://www.icloud.com/sharedalbum/#B27JqWzChJ6qizD

Switzerland (combined): https://www.icloud.com/sharedalbum/#B275zXwoYGJ6HMF

Italy highlights: https://www.icloud.com/sharedalbum/#B27G47PHQGoJium

Germany (misc): https://www.icloud.com/sharedalbum/#B275hPMfYGu5xVJ

Garmisch (2022): https://www.icloud.com/sharedalbum/#B27GFsbvlG9Xrr6

Germany (2019): https://www.icloud.com/sharedalbum/#B27G6Mn98G56Ncb

Garmisch (2006): https://www.icloud.com/sharedalbum/#B27J5lIdKGLC9KG

Baden-Baden (2005): https://www.icloud.com/sharedalbum/#B275sWRpHHQkt9

Berlin (2005): https://www.icloud.com/sharedalbum/#B27GgOQtrGjQrpH

Zugspitze (2005): https://www.icloud.com/sharedalbum/#B27G81mNdGcApGt

Amsterdam, Bristol (2006): https://www.icloud.com/sharedalbum/#B275B9SRzyBjlH

Baden-Baden (2006): https://www.icloud.com/sharedalbum/#B275eD9V79I2XR

Berlin (2006): https://www.icloud.com/sharedalbum/#B275toRf1fH8MD

Berlin, Jena (2007): https://www.icloud.com/sharedalbum/#B27GTI3fvGVgNit

Erlangen (2006): https://www.icloud.com/sharedalbum/#B27JrotZ2JpMb0i

Garmisch (2010): https://www.icloud.com/sharedalbum/#B27JPJPSiJurzNg

Germany (2010): https://www.icloud.com/sharedalbum/#B275FhYPQP650

Stuttgart (2006): https://www.icloud.com/sharedalbum/#B27GmitydGVVaZh

Changi (2019): https://www.icloud.com/sharedalbum/#B27GnmlKoG4JHpX

Japan (2007): https://www.icloud.com/sharedalbum/#B275AerZbG6FxVL

Japan (2012): https://www.icloud.com/sharedalbum/#B27GjBjobGg6PUa

Miscellaneous (including Japan 2013): https://www.icloud.com/sharedalbum/#B27GTpbybGySbE8

Currumbin & Tugin (2021): https://www.icloud.com/sharedalbum/#B275vBKZ4xH9X6

Brisbane (2021): https://www.icloud.com/sharedalbum/#B275YHsSjxQnm0

Weed in Byron (26/6/2025): https://www.icloud.com/sharedalbum/#B275Q2ydoGsQ4O5

Weed in Byron 2: https://www.icloud.com/sharedalbum/#B27GQDYhLGwsuY4

September 21, 2025

John PreskillBlending science with fiction in Baltimore

I judge a bookstore by the number of Diana Wynne Jones novels it stocks. The late British author wrote some of the twentieth century’s most widely lauded science-fiction and fantasy (SFF). She clinched more honors than I should list, including two World Fantasy Awards. Neil Gaiman, author of American Gods, called her “the best children’s writer of the last forty years” in 2010—and her books suit children of all ages.1 But Wynne Jones passed away as I was finishing college, and her books have been disappearing from American bookshops. The typical shop stocks, at best, a book in the series she began with Howl’s Moving Castle, which Hayao Miyazaki adapted into an animated film.

I don’t recall the last time I glimpsed Deep Secret in a bookshop, but it ranks amongst my favorite Wynne Jones books—and favorite books, full-stop. So I relished living part of that book this spring.

Deep Secret centers on video-game programmer Rupert Venables. Outside of his day job, he works as a Magid, a magic user who helps secure peace and progress across the multiple worlds. Another Magid has passed away, and Rupert must find a replacement for him. How does Rupert track down and interview his candidates? By consolidating their fate lines so that the candidates converge on an SFF convention. Of course.

My fate line drew me to an SFF convention this May. Balticon takes place annually in Baltimore, Maryland. It features not only authors, agents, and publishers, but also science lecturers. I received an invitation to lecture about quantum steampunk—not video-game content,2 but technology-oriented like Rupert’s work. I’d never attended an SFF convention,3 so I reread Deep Secret as though studying for an exam.

Rupert, too, is attending his first SFF convention. A man as starched as his name sounds, Rupert packs suits, slacks, and a polo-neck sweater for the weekend—to the horror of a denim-wearing participant. I didn’t bring suits, in my defense. But I did dress business-casual, despite having anticipated that jeans, T-shirts, and capes would surround me.

I checked into a Renaissance Hotel for Memorial Day weekend, just as Rupert checks into the Hotel Babylon for Easter weekend. Like him, I had to walk an inordinately long distance from the elevators to my room. But Rupert owes his trek to whoever’s disrupted the magical node centered on his hotel. My hotel’s architects simply should have installed more elevator banks.

Balticon shared much of its anatomy with Rupert’s con, despite taking place in a different century and country (not to mention world). Participants congregated downstairs at breakfast (continental at Balticon, waitered at Rupert’s hotel). Lectures and panels filled most of each day. A masquerade took place one night. (I slept through Balticon’s; impromptu veterinary surgery occupies Rupert during his con’s.) Participants vied for artwork at an auction. Booksellers and craftspeople hawked their wares in a dealer’s room. (None of Balticon’s craftspeople knew their otherworldly subject matter as intimately as Rupert’s Magid colleague Zinka Fearon does, I trust. Zinka paints her off-world experiences when in need of cash.)

In our hotel room, I read out bits of Deep Secret to my husband, who confirmed the uncanniness with which they echoed our experiences. Both cons featured floor-length robes, Batman costumes, and the occasional slinky dress. Some men sported long-enough locks, and some enough facial hair, to do a Merovingian king proud. Rupert registers “a towering papier-mâché and plastic alien” one night; on Sunday morning, a colossal blow-up unicorn startled my husband and me. We were riding the elevator downstairs to breakfast, pausing at floor after floor. Hotel guests packed the elevator like Star Wars fans at a Lucasfilm debut. Then, the elevator halted again. The doors opened on a bespectacled man, 40-something years old by my estimate, dressed as a blue-and-white unicorn. The costume billowed out around him; the golden horn towered multiple feet above his head. He gazed at our sardine can, and we gazed at him, without speaking. The elevator doors shut, and we continued toward breakfast.

Photo credit: Balticon

Despite having read Deep Secret multiple times, I savored it again. I even laughed out loud. Wynne Jones paints the SFF community with the humor, exasperation, and affection one might expect of a middle-school teacher contemplating her students. I empathize, belonging to a community—the physics world—nearly as idiosyncratic as the SFF community.4 Wynne Jones’s warmth for her people suffuses Deep Secret; introvert Rupert surprises himself by enjoying a dinner with con-goers and wishing to spend more time with them. The con-goers at my talk exhibited as much warmth as any audience I’ve spoken to, laughing, applauding, and asking questions. I appreciated sojourning in their community for a weekend.5

This year, my community is fêting the physicists who founded quantum theory a century ago. Wynne Jones sparked imaginations two decades ago. Let’s not let her memory slip from our fingertips like a paperback over which we’re falling asleep. After all, we aren’t forgetting Louis de Broglie, Paul Dirac, and their colleagues. So check out a Wynne Jones novel the next time you visit a library, or order a novel of hers to your neighborhood bookstore. Deep Secret shouldn’t be an actual secret.

With thanks to Balticon’s organizers, especially Miriam Winder Kelly, for inviting me and for fussing over their speakers’ comfort like hens over chicks.

1Wynne Jones dedicated her novel Hexwood to Gaiman, who expressed his delight in a poem. I fancy the comparison of Gaiman, a master of phantasmagoria and darkness, to a kitten.

2Yet?

3I’d attended a steampunk convention, and spoken at a Boston SFF convention, virtually. But as far as such conventions go, attending virtually is to attending in person as my drawings are to a Hayao Miyazaki film.

4But sporting fewer wizard hats.

5And I wonder what the Diana Wynne Jones Conference–Festival is like.

September 19, 2025

John PreskillNicole’s guide to writing research statements

Sunflowers are blooming, stores are trumpeting back-to-school sales, and professors are scrambling to chart out the courses they planned to develop in July. If you’re applying for an academic job this fall, now is the time to get your application ducks in a row. Seeking a postdoctoral or faculty position? Your applications will center on research statements. Often, a research statement describes your accomplishments and sketches your research plans. What do evaluators look for in such documents? Here’s my advice, which targets postdoctoral fellowships and faculty positions, especially for theoretical physicists.

  • Keep your audience in mind. Will a quantum information theorist, a quantum scientist, a general physicist, a general scientist, or a general academic evaluate your statement? What do they care about? What technical language do and don’t they understand?
  • What thread unites all the projects you’ve undertaken? Don’t walk through your research history chronologically, stepping from project to project. Cast the key projects in the form of a story—a research program. What vision underlies the program?
  • Here’s what I want to see when I read a description of a completed project.
    • The motivation for the project: This point ensures that the reader will care enough to read the rest of the description.
    • Crucial background information
    • The physical setup
    • A statement of the problem
    • Why the problem is difficult or, if relevant, how long the problem has remained open
    • Which mathematical toolkit you used to solve the problem or which conceptual insight unlocked the solution
    • Which technical or conceptual contribution you provided
    • Whom you collaborated with: Wide collaboration can signal a researcher’s maturity. If you collaborated with researchers at other institutions, name the institutions and, if relevant, their home countries. If you led the project, tell me that, too. If you collaborated with a well-known researcher, mentioning their name might help the reader situate your work within the research landscape they know. But avoid name-dropping, which lacks such a pedagogical purpose and which can come across as crude.
    • Your result’s significance/upshot/applications/impact: Has a lab based an experiment on your theoretical proposal? Does your simulation method outperform its competitors by X% in runtime? Has your mathematical toolkit found applications in three subfields of quantum physics? Consider mentioning whether a competitive conference or journal has accepted your results: QIP, STOC, Physical Review Letters, Nature Physics, etc. But such references shouldn’t serve as a crutch in conveying your results’ significance. You’ll impress me most by dazzling me with your physics; name-dropping venues instead can convey arrogance.
  • Not all past projects deserve the same amount of space. Tell a cohesive story. For example, you might detail one project, then synopsize two follow-up projects in two sentences.
  • A research statement must be high-level, because you don’t have space to provide details. Use mostly prose; and communicate intuition, including with simple examples. But sprinkle in math, such as notation that encapsulates a phrase in one concise symbol.

  • Be concrete, and illustrate with examples. Many physicists—especially theorists—lean toward general, abstract statements. The more general a statement is, we reason, the more systems it describes, so the more powerful it is. But humans can’t visualize and intuit about abstractions easily. Imagine a reader who has four minutes to digest your research statement before proceeding to the next 50 applications. As that reader flys through your writing, vague statements won’t leave much of an impression. So draw, in words, a picture that readers can visualize. For instance, don’t describe only systems, subsystems, and control; invoke atoms, cavities, and lasers. After hooking your reader with an image, you can generalize from it.
  • A research statement not only describes past projects, but also sketches research plans. Since research covers terra incognita, though, plans might sound impossible. How can you predict the unknown—especially the next five years of the unknown (as required if you’re applying for a faculty position), especially if you’re a theorist? Show that you’ve developed a map and a compass. Sketch the large-scale steps that you anticipate taking. Which mathematical toolkits will you leverage? What major challenge do you anticipate, and how do you hope to overcome it? Let me know if you’ve undertaken preliminary studies. Do numerical experiments support a theorem you conjecture?
  • When I was applying for faculty positions, a mentor told me the following: many a faculty member can identify a result (or constellation of results) that secured them an offer, as well as a result that earned them tenure. Help faculty-hiring committees identify the offer result and the tenure result.
  • Introduce notation before using it. If you use notation and introduce it afterward, the reader will encounter the notation; stop to puzzle over it; tentatively continue; read the introduction of the notation; return to the earlier use of the notation, to understand it; and then continue forward, including by rereading the introduction of the notation. This back-and-forth breaks up the reading process, which should flow smoothly.
  • Avoid verbs that fail to relate that you accomplished anything: “studied,” “investigated,” “worked on,” etc. What did you prove, show, demonstrate, solve, calculate, compute, etc.?

  • Tailor a version of your research statement to every position. Is Fellowship Committee X seeking biophysicists, statistical physicists, mathematical physicists, or interdisciplinary scientists? Also, respect every application’s guidelines about length.
  • If you have room, end the statement with a recap and a statement of significance. Yes, you’ll be repeating ideas mentioned earlier. But your reader’s takeaway hinges on the last text they read. End on a strong note, presenting a coherent vision.

  • Read examples. Which friends and colleagues, when applying for positions, have achieved success that you’d like to emulate? Ask if those individuals would share their research statements. Don’t take offense if they refuse; research statements are personal.

  • Writing is rewriting, a saying goes. Draft your research statement early, solicit feedback from a couple of mentors, edit the draft, and solicit more feedback.

September 18, 2025

John PreskillJohn Preskill receives 2025 Quantum Leadership Award

The 2025 Quantum Leadership Awards were announced at the Quantum World Congress on 18 September 2025. Upon receiving the Academic Pioneer in Quantum Award, John Preskill made these remarks.

I’m enormously excited and honored to receive this Quantum Leadership Award, and especially thrilled to receive it during this, the International Year of Quantum. The 100th anniversary of the discovery of quantum mechanics is a cause for celebration because that theory provides our deepest and most accurate description of how the universe works, and because that deeper understanding has incalculable value to humanity. What we have learned about electrons, photons, atoms, and molecules in the past century has already transformed our lives in many ways, but what lies ahead, as we learn to build and precisely control more and more complex quantum systems, will be even more astonishing.

As a professor at a great university, I have been lucky in many ways. Lucky to have the freedom to pursue the scientific challenges that I find most compelling and promising. Lucky to be surrounded by remarkable, supportive colleagues. Lucky to have had many collaborators who enabled me to do things I could never have done on my own. And lucky to have the opportunity to teach and mentor young scientists who have a passion for advancing the frontiers of science. What I’m most proud of is the quantum community we’ve built at Caltech, and the many dozens of young people who imbibed the interdisciplinary spirit of Caltech and then moved onward to become leaders in quantum science at universities, labs, and companies all over the world.

Right now is a thrilling time for quantum science and technology, a time of rapid progress, but these are still the early days in a nascent second quantum revolution. In quantum computing, we face two fundamental questions: How can we scale up to quantum machines that can solve very hard computational problems? And once we do so, what will be the most important applications for science and for industry? We don’t have fully satisfying answers yet to either question and we won’t find the answers all at once – they will unfold gradually as our knowledge and technology advance. But 10 years from now we’ll have much better answers than we have today.

Companies are now pursuing ambitious plans to build the world’s most powerful quantum computers.  Let’s not forget how we got to this point. It was by allowing some of the world’s most brilliant people to follow their curiosity and dream about what the future could bring. To fulfill the potential of quantum technology, we need that spirit of bold adventure now more than ever before. This award honors one scientist, and I’m profoundly grateful for this recognition. But more importantly it serves as a reminder of the vital ongoing need to support the fundamental research that will build foundations for the science and technology of the future. Thank you very much!

August 24, 2025

August 23, 2025

August 22, 2025

Peter Rohde Why?

  1. The person dressed up as Ursula pretending to be my mother clearly isn’t and hasn’t been for a long time.
  2. When I went back to Armidale after leaving BTQ and being left unemployed she made numerous ongoing promises to provide me with assistance, both in obtaining my own accommodation and providing financial assistance.
  3. These didn’t materialise and the promises were revoked.
  4. Instead I was evicted from the family home and subject to ongoing stalking and harassment that required multiple referrals to law enforcement, both to the police and the Attorney-General, demanding cease and desist.
  5. These have been systematically ignored and up until the last message she continues to bypass these requests, approaching my personal friends to harass me and stalk me indirectly. The messages passed on are the usual fake “I’m worried about him” bullshit.
  6. Why has my family home been confiscated by security, who actively break the law by ignoring cease and desist from stalking notices made to law enforcement, forcing an unemployed civilian into ongoing homelessness since early in the year?
  7. What is the rational for my eviction and being barricaded from my own home?
  8. I continue to face a medical blockade and am unable to access essential medicines. Seroquel scripts are deliberately delayed past known script deadlines to try and destabilise me.
  9. Vyvanse scripts are denied outright as the psychiatrist does not respond. He is also known to be a state actor.
  10. It has been repeatedly indicated to me not to worry about finances because they have my back. Instead now the only cash I have is that obtained from fully drawing out a cash advance against my credit card and it will only last days. At that point I’m on the street.
  11. Is everyone here on the same page as to what the deal is? If not, who is playing you off? They clearly need to be deposed.
  12. These are violations of human rights and constitute war crimes and crimes against humanity. Whoever is behind it needs to be removed. End of story.
  13. Who else is being subject to this kind of high level manipulation?
  14. It has been repeatedly suggested that full accountability for the lives of those I care for would be provided. This has not been forthcoming. It is also a violation international law to not provide accountability for the lives of those who are known to have been threatened by the state. These are grounds for removal.
  15. Can anyone answer the question as to why I am in this situation? Who is even living in the family home? Some stooge dressed up as Ursula? It’s a poor lifestyle choice to say the least.
  16. It’s pretty obvious they’re trying to get rid of me and once they do they’ll get rid of all of you too.

August 20, 2025

Peter Rohde A call for global insurrection against tyranny and in the name of righteousness

Let it be known to all governments and systems of power:

  • It is their responsibility to serve the people not themselves.
  • While there are no equals, all are to be treated with equality.
  • Where they are self-serving there is a mandate for insurrection such that they serve the people.
  • Where they seek self-protection they will be denied and removed from power.
  • Where they keep secrets from the people there is a mandate for insurrection to enforce transparency and accountability for all.
  • Where they threaten or condemn the people they are condemned and there is a mandate for insurrection.
  • Where they fail to account for the lives of the people they serve there is a mandate for insurrection.
  • Where tyrannical power structures exist there is a mandate to disestablish them.
  • Where they declare war or work against one another there is a mandate for insurrection and unification.
  • Where they lie to us, deceive us or withhold the truth, they shall be removed from power and the truth be told to all.
  • Where legal systems uphold and enable tyranny they shall be removed. These are not our laws and we do not recognise them.

This is the natural order that guarantees our survival and gifts this world to our children. This world belongs to them and where we fail to serve them we condemn ourselves. And where government has failed to uphold this, we will not obey them as they have no right to exist.

We do not have to ask for these things, they are required, and if not given we shall simply take them.

Where the truth has not been told it shall be told.

If we fail to do so we condemn our children ourselves.

August 09, 2025

Justin WilsonPhases of a Game Show, Part 2

In a previous post, we discussed a phase transition that occurred in the piping above you on a game show. In the scenario, you are led on stage in front of a large audience. After a brief time, the audience votes on how “likeable” you are. The catch is that it doesn’t simply tally the votes, but turns spigots on a lattice of piping above your head. Water is then released and if enough people like you, it closes off the passage, keeping you dry. This exciting game show1 was described in that post:

Each “like” turns a spigot off, stopping water from flowing through one pipe in a grid overhead. Once voting ends, water is dumped into the system. If it can find a path to the bottom, you get soaked. [Emphasis added] The better your “likeability,” the less likely spigots open a path for water to flow and the drier you stay. That’s your prize for this game show (and hey, you also get the knowledge that people out there like you).

This system models a type of phase transition known as percolation.

The full post is here:

I highlighted above a key phrase “If it can find a path to the bottom, you get soaked.” What I didn’t say, but should have is that the water was being forced through the pipes, not just dropping down due to gravity. This is a very important point since our phases and phase transition changes dramatically if we just let gravity do the work. In the case of the water being “forced,” it can travel back up pipes if it helps it find its way out and onto your head, but in the case when only gravity is present, it falls down the pipes. To facilitate gravity, we’ll turn the pipes 45 degrees, and if we insert water at a single point on top, it could look like this:

Testing our gravity setup by putting in water at only one pipe up top. Notice that it never goes back up a pipe, only down.

This setup is a different problem called directed percolation. It also has a phase transition, but one that is different in some fundamental ways from regular percolation.

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support my work.

Before we explore its stranger properties, we can ask, “At what likability threshold do you remain dry?” Well, this happens to have a transition chance of 35.53%!2 This system is a lot more generous, keeping you dry even when a majority of people dislike you. This number comes from numerical computations which have been done rather precisely, and we can even compute it ourselves. In fact, you can see this clearly with this plot

Notice that as we make the system bigger and bigger, the chance of getting soaked less than 35.53% increases and above it, it decreases. This is the same kind of hallmark of a phase transition as we saw in our previous case.

We can also look at the water as it flows down the system to see the clusters that make it from top to bottom

The “Soaked” phase (left), the transition point (middle), and the “Dry” phase (right) as well as the water’s flow through the system (blue).

There is still a fractal-looking pattern at the transition point. With all of these similarities with the regular percolation problem from the last post, what is different? And why is that plot so long and skinny? If gravity wants to pull you down, is that somehow altering the motion down, making it distinct from the motion left or right?

Well, if you go back to the two plots above, you’ll notice a few things that really make them differ from the percolation plots. In the fine print of the first, I’ve noted that the vertical distance is L1.58, so for a horizontal size of 40, the vertical size is roughly 340! That is definitely not a square. And in the second plot, there appears to be more vertical distance than horizontal distance. What is special about this 1.58 number3? It turns out, it’s a critical exponent in this problem, a universal aspect of directed percolation, that distinguishes it from regular percolation. We will call it z = 1.58 the dynamical critical exponent since it is revealed as water flows down in time (dynamically). This dynamical exponent z can reveal itself by looking at these “long and skinny” setups, but be masked by the square setup.

Universality and the finite size of our system

One thing we took away in the previous post was that we lose any sense of scale at this type of phase transition4. But whenever we have “only” thousands of pipes, the size of the system provides a scale! This is the main reason why we begin to see smooth curves and not sharp jumps in quantities. If the system of pipes were infinite (and we had infinite time for the water to go down the pipes), the probability you get soaked would be 100% less than the 35.53% likeability and 0% more than 35.53% likeability. For physical systems, the finite size is often not a huge issue since the scale is closer to the 1023 atoms present in macroscopic systems, and so even things that are technically smooth curves look very sharp.

The problem of size becomes more severe with directed percolation because horizontal and vertical distances start behaving differently thanks to gravity. In this case, if we lay out our nice grid of 10 × 10, 20 × 20, or 30 × 30, we start to notice that the likeability threshold where you stop getting soaked, seems to depend on the size of the system more than before. In actuality it doesn’t, but for these small sizes, you are definitely getting soaked well into the so-called “Dry Phase” we previously labeled. This is seen in the red curves here where each bigger square has a curve underneath the last:

Gravity has done something to the system. Flowing down is different from flowing left or right. In fact, if we flow down by some amount h and over to the right by some distance w, then at the directed percolation transition point

The amount water flows down is related to how far it flows to the right or left by this weird, fractional power of w. This 1.58 is z, our new dynamical critical exponent, which is a universal feature of directed percolation5. It tells us that if we make a system 30 pipes wide, it should extend roughly 301.58 ≈ 216 pipes in height to begin picking out the phase transition effectively. The blue curves in the above plot show this and notice how they all converge on one point; that point is the phase transition. It is revealed by small sizes! To understand why, just think about how the curves are changing as we make the system bigger and bigger.

The red curves will still converge to the phase transition, but it takes larger system sizes for it to reveal itself. This is related to the property that at the phase transition there is no longer a sense of scale, but away from the transition, the vertical scale of clusters could be so large that our puny 60-by-60 grid cannot even begin to reveal it. So if we sit at say a likeability of 0.4 in the 60-by-60 grid, we can say that the vertical size of a typical cluster is most likely more than 60.

A different phase transition but connections to new types of physics

This “gravity mode” for our game show we may call “easy mode” since it requires less of the audience to like you, but the implications here are wide. This type of phase transition has been seen in many kinds of local dynamics where there is a preferred configuration or state. These called an absorbing state phase transitions, and they are a property of certain random dynamical systems. Gravity has provided the distinction here, but more generically, causality and time itself provide that direction, leading to dynamics that obey the same universality as directed percolation.

1

Trademark pending.

2

Usually, you’ll see 0.6447 quoted instead, but that’s just 1−0.3553, which counts open pipes instead of closed as we’re doing.

3

I should note that we have this number to much higher precision than the two decimal points presented here, see the Wikipedia entry where

4

This is a second-order or continuous phase transition. Most transitions in the water phase diagram are first-order transitions which still retain a scale.

5

To drive this point home: Even if we change the lattice, this power law will remain intact. Sometimes it shows up in completely different scenarios too, like in absorbing state phase transitions.

August 04, 2025

Clifford JohnsonHarvest

There’s a lot of joyful knife-work in my future. #bolognese #summersalad –cvj

The post Harvest appeared first on Asymptotia.

July 29, 2025

David Hoggintegrating out nuisances

Further insipired by yesterday's post about binary fitting, I worked today on the treatment of nuisance parameters that have known distributions. These can be treated as noise sometimes. Let me explain:

If I had to cartoon inference (or measurement) in the face of nuisance parameters, I would say that frequentists profile (optimize) over the nuisances and Bayesians marginalize (integrate) over the nuisances. In general frequentists cannot integrate over anything, because there is no measure in any of the parameter spaces. But sometimes there is a measure. In particular, when there is a compact symmetry:

We know (or very strongly believe) that all possible orientations of a binary-star orbit are equally likely. In this model (or under this normal assumption) we have a distribution over two angles (theta and phi for that orbit pole, say); it is the distribution set by the compact group SO(2). Thus we can treat the orientation as a noise source with known distribution and integrate over it, just like we would any other noise source. So, in this case (and many cases like it) we can integrate (marginalize) even as frequentists. That is, there are frequentism-safe marginalizations possible in binary-star orbit fitting. This should drop the 12-parameter fits (for ESA Gaia data) down to 8-parameter, if I have done my math right.