Planet Musings

September 15, 2025

Terence TaoSmooth numbers and max-entropy

Given a threshold {y>1}, a {y}-smooth number (or {y}-friable number) is a natural number {n} whose prime factors are all at most {y}. We use {\Psi(x,y)} to denote the number of {y}-smooth numbers up to {x}. In studying the asymptotic behavior of {\Psi(x,y)}, it is customary to write {y} as {x^{1/u}} (or {x} as {y^u}) for some {u>0}. For small values of {u}, the behavior is straightforward: for instance if {0 < u < 1}, then all numbers up to {x} are automatically {y}-smooth, so

\displaystyle  \Psi(x,y) \sim x

in this case. If {1 < u < 2}, the only numbers up to {x} that are not {y}-smooth are the multiples of primes {p} between {y} and {x}, so

\displaystyle  \Psi(x,y) \sim x - \sum_{y < p \leq x} (\frac{x}{p} + O(1))

\displaystyle  \sim x - x (\log\log x - \log\log y)

\displaystyle  \sim x (1 - \log u)

where we have employed Mertens’ second theorem. For {2 < u < 3}, there is an additional correction coming from multiples of two primes between {y} and {x}; a straightforward inclusion-exclusion argument (which we omit here) eventually gives

\displaystyle  \Psi(x,y) \sim x (1 - \log u + \int_2^u \frac{\log(t-1)}{t} dt)

in this case.

More generally, for any fixed {u>0}, de Bruijn showed that

\displaystyle  \Psi(x, y) \sim \rho(u) x

where {\rho(u)} is the Dickman function. This function is a piecewise smooth, decreasing function of {u}, defined by the delay differential equation

\displaystyle  u \rho'(u) + \rho(u-1) = 0

with initial condition {\rho(u) = 1} for {0 \leq u \leq 1}.

The asymptotic behavior of {\rho(u)} as {u \rightarrow \infty} is rather complicated. Very roughly speaking, it has inverse factorial behavior; there is a general upper bound {\rho(u) \leq 1/\Gamma(u+1)}, and a crude asymptotic

\displaystyle  \rho(u) = u^{-u+o(u)} = \exp( - u \log u + o(u \log u)).

With a more careful analysis one can refine this to

\displaystyle  \rho(u) = \exp( - u \log u - u \log\log u + u + o(u)); \ \ \ \ \ (1)

and with a very careful application of the Laplace inversion formula one can in fact show that

\displaystyle  \rho(u) \sim \sqrt{\frac{\xi'(u)}{2\pi}} \exp( \gamma - u \xi(u) + \int_0^{\xi(u)} \frac{e^s - 1}{s} ds) \ \ \ \ \ (2)

where {\gamma} is the Euler-Mascheroni constant and {\xi(u)} is defined implicitly by the equation

\displaystyle  e^{\xi(u)} - 1 = u \xi(u). \ \ \ \ \ (3)

One cannot write {\xi(u)} in closed form using elementary functions, but one can express it in terms of the Lambert {W} function as {\xi(u) = -W(-\frac{1}{u} e^{-1/u}) - 1/u}. This is not a particularly enlightening expression, though. A more productive approach is to work with approximations. It is not hard to get the initial approximation {\xi(u) \approx \log u} for large {u}, which can then be re-inserted back into (3) to obtain the more accurate approximation

\displaystyle  \xi(u) = \log u + \log\log u + O(1)

and inserted once again to obtain the refinement

\displaystyle  \xi(u) = \log u + \log\log u + O(\frac{\log\log u}{\log u}).

We can now see that (2) is consistent with previous asymptotics such as (1), after comparing the integral {\int_0^{\xi(u)} \frac{e^s - 1}{s} ds} to

\displaystyle  \int_0^{\xi(u)} \frac{e^s - 1}{\xi(u)} ds = u - 1.

For more details of these results, one can see for instance this survey by Granville.

This asymptotic (2) is quite complicated, and so one does not expect there to be any simple argument that could recover it without extensive computation. However, it turns out that one can use a “maximum entropy” to get a reasonably good heuristic approximation to (2), that at least reveals the role of the mysterious function {\xi(u)}. The purpose of this blog post is to give this heuristic.

Viewing {x = y^u}, the task is to try to count the number of {y}-smooth numbers of magnitude {y^u}. We will propose a probabilistic model to generate {y}-smooth numbers as follows: for each prime {p \leq y}, select the prime {p} with an independent probability {c_p/p} for some coefficient {c_p}, and then multiply all the selected primes together. This will clearly generate a random {y}-smooth number {n}, and by the law of large numbers, the (log-)magnitude of this number should be approximately

\displaystyle  \log n \approx \sum_{p \leq y} \frac{c_p}{p} \log p, \ \ \ \ \ (4)

(where we will be vague about what “{\approx}” means here), so to obtain a number of magnitude about {y^u}, we should impose the constraint

\displaystyle  \sum_{p \leq y} \frac{c_p}{p} \log p = u \log y. \ \ \ \ \ (5)

The indicator {1_{p|n}} of the event that {p} divides this number is a Bernoulli random variable with mean {c_p/p}, so the Shannon entropy of this random variable is

\displaystyle  \mathbf{H}(1_{p|n}) = - \frac{c_p}{p} \log(\frac{c_p}{p}) - (1 - \frac{c_p}{p}) \log(1 - \frac{c_p}{p}).

If {c_p} is not too large, then Taylor expansion gives the approximation

\displaystyle  \mathbf{H}(1_{p|n}) \approx \frac{c_p}{p} \log p - \frac{c_p}{p} \log c_p + \frac{c_p}{p}.

Because of independence, the total entropy of this random variable {n} is

\displaystyle  \mathbf{H}(n) = \sum_{p \leq y} \mathbf{H}(1_{p|n});

inserting the previous approximation as well as (5), we obtain the heuristic approximation

\displaystyle  \mathbf{H}(n) \approx u \log y - \sum_{p \leq y} \frac{c_p}{p} (\log c_p - 1).

The asymptotic equipartition property of entropy, relating entropy to microstates, then suggests that the set of numbers {n} that are typically generated by this random process should be approximately

\displaystyle  \exp(\mathbf{H}(n)) \approx \exp\left(u \log y - \sum_{p \leq y} \frac{c_p}{p} (\log c_p - 1)\right)

\displaystyle  = \exp\left(- \sum_{p \leq y} \frac{c_p \log c_p - c_p}{p}\right) x.

Using the principle of maximum entropy, one is now led to the approximation

\displaystyle  \rho(u) \approx \exp\left(- \sum_{p \leq y} \frac{c_p \log c_p - c_p}{p}\right). \ \ \ \ \ (6)

where the weights {c_p} are chosen to maximize the right-hand side subject to the constraint (5).

One could solve this constrained optimization problem directly using Lagrange multipliers, but we simplify things a bit by passing to a continuous limit. We take a continuous ansatz {c_p = f(\log p / \log y)}, where {f: [0,1] \rightarrow {\bf R}} is a smooth function. Using Mertens’ theorem, the constraint (5) then heuristically becomes

\displaystyle  \int_0^1 f(t)\ dt = u \ \ \ \ \ (7)

and the expression (6) simplifies to

\displaystyle  \rho(u) \approx \exp( - \int_0^1 \frac{f(t) \log f(t) - f(t)}{t}\ dt). \ \ \ \ \ (8)

So the entropy maximization problem has now been reduced to the problem of minimizing the functional {\int_0^1 \frac{f(t) \log f(t) - f(t)}{t}\ dt} subject to the constraint (7). The astute reader may notice that the integral in (8) might diverge at {t=0}, but we shall ignore this technicality for the sake of the heuristic arguments.

This is a standard calculus of variations problem. The Euler-Lagrange equation for this problem can be easily worked out to be

\displaystyle  \frac{\log f(t)}{t} = \lambda

for some Lagrange multiplier {\lambda}; in other words, the optimal {f} should have an exponential form {f(t)= e^{\lambda t}}. The constraint (7) then becomes

\displaystyle  \frac{e^\lambda - 1}{\lambda} = u

and so the Lagrange multiplier {\lambda} is precisely the mysterious quantity {\xi(u)} appearing in (2)! The formula (8) can now be evaluated as

\displaystyle  \rho(u) \approx \exp\left( - \int_0^1 \frac{e^{\xi(u) t} \xi(u) t - e^{\xi(u) t}}{t}\ dt \right)

\displaystyle  \approx \exp\left( - e^{\xi(u)} + 1 + \int_0^1 \frac{e^{\xi(u) t} - 1}{t}\ dt + \int_0^1 \frac{1}{t}\ dt \right)

\displaystyle  \approx \exp\left( - u \xi(u) + \int_0^{\xi(u)} \frac{e^s - 1}{s}\ ds + C\right)

where {C} is the divergent constant

\displaystyle  C = \int_0^1 \frac{1}{t}\ dt.

This recovers a large fraction of (2)! It is not completely accurate for multiple reasons. One is that the hypothesis of joint independence on the events {p|n} is unrealistic when trying to confine {n} to a single scale {x}; this comes down ultimately to the subtle differences between the Poisson and Poisson-Dirichlet processes, as discussed in this previous blog post, and is also responsible for the otherwise mysterious {e^\gamma} factor in Mertens’ third theorem; it also morally explains the presence of the same {e^\gamma} factor in (2). A related issue is that the law of large numbers (4) is not exact, but admits gaussian fluctuations as per the central limit theorem; morally, this is the main cause of the {\sqrt{\frac{\xi'(u)}{2\pi}}} prefactor in (2).

Nevertheless, this demonstrates that the maximum entropy method can achieve a reasonably good heuristic understanding of smooth numbers. In fact we also gain some insight into the “anatomy of integers” of such numbers: the above analysis suggests that a typical {y}-smooth number {n} will be divisible by a given prime {p \sim y^t} with probability about {e^{\xi(u) t}/p}. Thus, for {t = 1 - O(1/\log u)}, the probability of being divisible by {p} is elevated by a factor of about {\asymp e^{\xi(u)} \asymp u \log u} over the baseline probability {1/p} of an arbitrary (non-smooth) number being divisible by {p}; so (by Mertens’ theorem) a typical {y}-smooth number is actually largely comprised of something like {\asymp u} prime factors all of size about {y^{1-O(1/\log u)}}, with the smaller primes contributing a lower order factor. This is in marked contrast with the anatomy of a typical (non-smooth) number {n}, which typically has {O(1)} prime factors in each hyperdyadic scale {[\exp\exp(k), \exp\exp(k+1)]} in {[1,n]}, as per Mertens’ theorem.

September 14, 2025

Jordan EllenbergKlartag improves the sphere-packing constant

Hey! It escaped my notice until this week that Boaz Klartag has recently shown the existence of lattice sphere packings in dimension d of density at least c d^2 2^-d. This lower bound was stuck at d 2^-d for a long time, and was only recently improved by a factor of log d, which I learned about from Marcus Michelen‘s job talk (unfortunately for us, as the link will show you, he chose to go to Northwestern!) So it’s very cool that Klartag jumps this up a whole power of d, and by an entirely new method. As in Michelen’s work, though, the approach is probabilistic.

Here’s the idea in a nutshell. Showing the existence of a lattice packing with large density is equivalent to showing the existence of a large ellipsoid which contains no nontrivial point of the standard lattice. (Just take a linear transformation which sends the big ellipsoid to a unit sphere, and which thus has small determinant; apply it to the lattice too, and you have a dense lattice which misses the unit sphere, and Bob’s your uncle.) It turns out you can get such an ellipsoid by almost-aimless wandering! More precisely: start with some ellipsoid whose disjoint from the (nonzero elements of the) lattice, and then carry out some form of Brownian motion in the space of all ellipsoids, subject to one very important rule: every time your ellipse hits a lattice point x, you “freeze” that intersection point, and from now on the ellipsoid wanders in the one-dimension-smaller space of ellipsoids which pass through x. The process thus stops when you’ve hit about (1/2)n^2 points, because that uses up all your degrees of freedom. The game, then, is to show that there’s a slight constant drift towards increasing volume, so that in the length cn^2 lifetime of your process, the ellipsoid acquires c n^2 volume, and that’s exactly where the n^2 in the theorem comes from.

Very cool that such a charming idea actually works!

Jordan EllenbergBertrand Russell on rationality and math

“Aristotle, so far as I know, was the first man to proclaim explicitly that man is a rational animal. His reason for this view was one which does not now seem very impressive; it was, that some people can do sums.”

Russell was one quotable individual.

September 13, 2025

n-Category Café A Shadow of Triality?

It’s well known that you can construct the octonions using triality. One statement of triality is that Spin(8)Spin(8) has nontrivial outer automorphisms of order 3. On the other hand, the octonions have nontrivial inner automorphisms of order 3. My question: can we deduce one of these facts from the other?

The second fact is perhaps not very well known. It may even be hard to understand what it means. Though the octonions are nonassociative, for any nonzero octonion gg the map

f: 𝕆 𝕆 x gxg 1 \begin{array}{rccl} f \colon & \mathbb{O} &\to& \mathbb{O} \\ & x & \mapsto & g x g^{-1} \end{array}

is well-defined, since (gx)g 1=g(xg 1)(g x)g^{-1} = g(x g^{-1}), which one can show using the fact that the octonions are alternative. More surprisingly, whenever g 6=1g^6 = 1, this map ff is an automorphism of the octonions:

f(x+y)=f(x)+f(y),f(xy)=f(x)f(y)x,y𝕆 f(x+y) = f(x) + f(y) , \qquad f(x y) = f(x) f(y) \qquad \forall x,y \in \mathbb{O}

and ff has order 3:

f(f(f(x)))=xx𝕆 f(f(f(x))) = x \qquad \forall x \in \mathbb{O}

To understand this latter fact, we can look at

Theorem 2.1 here implies that an octonion gg with |g|=1{|g|} = 1 defines an inner automorphism f:xgxg 1f \colon x \to g x g^{-1} if and only if xx has order 6.

However, the result is stated differently there. Paraphrasing somewhat, Lamont’s theorem says that any g𝕆g \in \mathbb{O} that is not a real multiple of 1𝕆1 \in \mathbb{O} defines an inner automorphism f:xgxg 1f \colon x \to g x g^{-1} if and only if gg obeys

4Re(g) 2=|g| 2 4 \mathrm{Re}(g)^2 = {|g|}^2

This equation is equivalent to Re(g)=±12|g|\operatorname{Re}(g) = \pm \frac{1}{2} {|g|}, which is equivalent to gg lying at either a 60 60^\circ angle or a 120 120^\circ angle from the octonion 11.

Octonions on the real line clearly define inner automorphisms. Thus, a nonzero octonion gg defines an inner automorphism if and only if its angle from 11 is 0 0^\circ, 60 60^\circ, 120 120^\circ or 180 180^\circ. In this case we can normalize gg without changing the inner automorphism it defines, and then we have g 6=1g^6 = 1. Note also that gg and g-g define the same inner automorphism.

It follows that an octonion gg on the unit sphere defines an inner automorphism iff g 6=1g^6 = 1, and that every nontrivial inner automorphism of 𝕆\mathbb{O} has order 3.

However, if you look at Lamont’s proof, you’ll see the equation 4Re(g) 2=|g| 24 \operatorname{Re}(g)^2 = {|g|}^2 plays no direct role! Instead, he really uses the assumption that g 3g^3 is a real multiple of 11, which is implied by this equation (as easily shown using what we’ve just seen).

From Lamont’s work, one can see the Moufang identities and the characteristic equation for octonions are what force all inner automorphisms of the octonions to have order 3.

Thus, an argument giving a positive answer to my question might involve a link between triality and the Moufang identities. Conway and Smith seem to link them in On Quaternions and Octonions. But I haven’t figured out how to get from the outer automorphisms of Spin(8)\text{Spin}(8) to the inner automorphisms of 𝕆\mathbb{O}, or vice versa!

I asked about this on MathOverflow, but I thought some people here would also be interested.

n-Category Café Burrito Monads, Arrow Kitchens, and Freyd Category Recipes

Guest post by Khyathi Komalan and Andrew Krenz

From Lawvere’s Hegelian taco to Baez’s layer cake analogy to Eugenia Cheng’s How to Bake Pi, categorists have cultivated a rich tradition of culinary metaphors and similes. A well-known example in the world of computation is Mark Dominus’s “monads are like burritos” — where a tortilla (computational context) wraps diverse ingredients (values) to create a cohesive entity (effectful value) whose burrito structure is maintained as the meal moves down the assembly line (undergoes computations).

Monads, like burritos, come in many different varieties. In computer science monads serve to streamline computational patterns such as exception handling and context management. We illustrate these two examples by analogy.

Imagine you work at a burrito truck.

If a customer orders a burrito sans rice but rice is accidentally added, it can’t be served. The Maybe monad handles exceptions such as this — when something goes wrong, it returns a special “Nothing” value rather than a flawed result, and once a failure occurs, all subsequent steps automatically preserve this state avoiding the need for repetitive error-checking.


Diagram 1

Figure 1: The Maybe Monad illustrated with the burrito-making process


In Haskell, the parameterized type “Maybe a” has two constructors, “Just a” and “Nothing.” The former is an alias for values of type “a” whereas the latter is indicative of an error. The following Haskell code exhibits the maybe type as an instance of the monad class:

instance Maybe Monad where
return = Just
Nothing >>= f = Nothing
(Just x) >>= f = f x

the return function has type a -> Maybe a, which is suggestive of its role as the monad unit. The so-called bind operation >>= has type Maybe a -> (a -> Maybe b) -> Maybe b, and corresponds to a bare-bones Kleisli composition (see Monads: Programmer’s Definition for details).

A slight generalization allows for descriptive error messages.

Definition. Given a collection of exceptions EE, there is an associated Either monad (()+E,η,μ)((-)+E, \eta, \mu).

  • η X:XX+E\eta_X:X \to X + E is the coproduct insertion
  • μ X:X+E+EX+E\mu_X:X + E + E \to X + E collapses two copies of EE into one
  • Kleisli morphisms are computations that may fail XY+EX \to Y + E
  • Kleisli composition automatically propagates exceptions

Of course, either monads are simply maybe monads with a set in place of the constant/singleton “Nothing” and they allow us not only to say that an error has occured, but also to indicate what that error was.

Now suppose one of your regular customers walks up to the window and orders “the usual.” Luckily you’ve recorded their preferences in a recipe book. The act of following the appropriate recipe is akin to executing computations that depend on a global read-only state. The * Reader monad * is the functional programmer’s way of incorporating this impure concept in pure functional terms.

Diagram 2


Figure 2: The Reader Monad illustrated with the burrito-making process

Definition. Given a collection of environments EE, there is an associated Reader monad (() E,η,μ)((-)^E, \eta, \mu).

  • η X:XX E\eta_X : X \to X^E turns elements into constant functions xλe.xx \mapsto \lambda e. x
  • μ X:(X E) EX E\mu_X : (X^E)^E \to X^E turns function-valued functions into functions via diagonal evaluation fλe.f(e)(e)f \mapsto \lambda e. f(e)(e)
  • Kleisli morphisms convert inputs into executable functions from environments to outputs XY EX \to Y^E
  • Composition in the Kleisli category keeps track of the (immutable) environment as computations are chained together.

Here is the same definition given as an instance of the Haskell monad class:

instance Monad ((->) r) where
return x = \_ -> x
g >>= f = \e -> f (g e) e

The seminal paper of Moggi has several other interesting examples illustrating the power of monads. Nevertheless, monads may not always suffice for all of our needs. For example, what would happen if our burrito truck suddenly exploded in popularity requiring automation of repetative processes and parallel work stations?

This is where “Arrows” enter the picture. Introduced by John Hughes in 2000, Arrows generalize strong monads. Because of this, Arrows handle more complicated computational patterns in a natural way. While monads wrap values in computational contexts (like burritos in tortillas), Arrows can represent entire preparation processes capable of coordinating multiple inputs while maintaining awareness of the broader kitchen environment.

Arrows come with three core operations that determine their behaviour; looking at their types, we see that Arrows are evocative of a lax internal hom that interacts with binary products.

class Arrow a where
arr :: (x -> y) -> a x y
(>>>) :: a x y -> a y z -> a x z
first :: a x y -> a (x,z) (y,z)
  1. arr turns functions into “Arrows.” This is like incorporating a standard burrito recipe or preparation step into the food truck’s workflow — taking a simple instruction like “add beans, then cheese” and automating it within our kitchen’s setup.
  2. >>> composes composable Arrows. This allows for separately automated processes to be seamlessly strung together.
  3. first enacts an automated process on one burrito while simultaneously passing a second burrito through the station.

These data are subject to 9 axioms, which we eventually discuss below.

Diagram 3
Figure 3: Arrow Operations. The three fundamental operations of Arrows enable complex workflows beyond monadic structures.

Shortly before Arrows were introduced, Power, Robinson, and Thielecke were working on Freyd categories — a categorical structure designed to model “effectful” computation. Using our simile, a Freyd category formalizes the relationship between an ideal burrito recipe (pure theory) and the real-world process of making that burrito in a particular kitchen.

A Freyd category consists of three main components:

  1. A category CC with finite products which can be thought of as the syntax of our kitchen. In other words, CC is like a recipe book containing the abstract information one needs to interpret and implement in the context of an actual kitchen.
  2. A symmetric premonoidal category KK which plays the more semantic role of our real world kitchen.
  3. An identity-on-objects functor J:CKJ:C \to K which faithfully translates pure recipes into physical processes that work within the specific setup of the kitchen KK.

    Diagram 4
    Figure 4: Freyd Category Structure. The relationship between pure recipes (category C) and real-world kitchen operations (category K), connected by the identity-on-objects functor J that preserves structure while accommodating practical constraints.

Although Arrows originated in Haskell, a highly abstract functional programming language, researchers began noticing apparent correspondences between the components of Arrows and those of Freyd categories. These two structures, developed from different starting points, seemed to address the same fundamental challenge: how to systematically manage computations that involve effects, multiple inputs and outputs, and context-awareness. Therefore, it was hypothesized that Arrows are equivalent to Freyd categories.

As a part of the Adjoint School, our group has been focusing on R. Atkey’s work, which dispells this folklore and precisely formulates the relationship between Arrows and Freyd categories. Just as Atkey asks in the title of his paper, this blog post will investigate the question of “what is a categorical model of Arrows?” The answer not only clarifies the theoretical underpinnings of these structures, but also reveals practical insights for programming language design and quantum computation models. Ultimately, we will see that there are indeed subtle differences between Arrows and Freyd categories.


Key Insights: - Monads encapsulate computational effects by wrapping values in contexts, much like burritos wrap ingredients in tortillas - Different monads (Maybe, Reader, etc…) deal with different patterns like exception handling and context management - Arrows generalize monads to handle multiple inputs and coordinate complex processes, like managing an entire kitchen rather than just making individual burritos


Beyond the Kitchen: Arrows and Freyd Categories

Formally, a monad on a category CC is a monoid in the category of endofunctors of CC. Arrows, like monads, are monoids in a certain category of functors. To be more specific, the structure of an Arrow on a category CC can be described as a monoid in the category of strong profunctors on CC. Let’s take a closer look at this construction.

Arrows A profunctor PP on a category CC is a functor P:C op×CSet.P: C^{\text{op}}\times C \to \text{Set}. Intuitively, a profunctor associates to each pair of objects a set of “generalized morphisms” between those objects.

The identity profunctor is simply id(x,y):=C(x,y)\text{id}(x, y) := C(x, y), which uses the hom-sets of CC.

Composition of profunctors is defined as a coend. Given profunctors PP and QQ, their composition is the following profunctor:

(P*Q)(x,z)= yP(x,y)×Q(y,z)(P * Q)(x, z) = \int^y P(x, y) \times Q(y, z)

Notice that this formula is vaguely reminiscent of a dot product; replacing the integral with a sum over yy, and the cartesian product with multiplication, it looks like the dot product of the row vector P(x,)P(x,-) with the column vector Q(,z)Q(-,z).

operations. –> We will now unpack this data to reach a more down-to-earth description of Arrows. This resulting characterization aligns more closely with the way in which Arrows are implemented in programming languages like Haskell.

Definition. An Arrow in a cartesian closed category CC consists of a mapping on objects and three families of morphisms:

  • A mapping on objects Ar:ob(C)×ob(C)ob(C)\text{Ar} : \text{ob}(C) \times \text{ob}(C) \to \text{ob}(C)

This defines the Arrow type constructor, which takes input and output types and produces an Arrow type between them.

  • A family of morphisms arr:Y XAr(X,Y)\text{arr} : Y^X \to \text{Ar}(X, Y)

This operation lifts a pure function into the Arrow context, allowing regular functions to be treated as Arrows.

  • A family of morphisms :Ar(X,Y)×Ar(Y,Z)Ar(X,Z)\ggg : \text{Ar}(X, Y) \times \text{Ar}(Y, Z) \to \text{Ar}(X, Z)

    This enables sequential composition of Arrows, similar to function composition but now in terms of Arrows.

  • A family of morphisms first:Y XAr(X×W,Y×W)\text{first} : Y^X \to \text{Ar}(X \times W, Y \times W)

This is perhaps the most distinctive operation. Intuitively, it allows an Arrow to process the first component of a pair while leaving the second component unchanged.

These data are subject to nine axioms which govern their interactions. To make these abstract operations more concrete, consider the following example, where Ar(x,y):=Y X,\text{Ar}(x, y) := Y^X, arr:=id (Y X),\text{arr} := \text{id}_{(Y^X)}, :=composition,\ggg := \text{composition}, and first(f):=f×id.\text{first}(f) := f \times \text{id}. In what follows we list the Arrow laws and draw commutative diagrams based on this example.

The Arrow laws arr(id)a=a\text{arr}(\text{id})\ggg a=a and aarr(id)=aa\ggg \text{arr}(\text{id})=a express left and right unitality of identities under composition.

Diagram 5
Figure 5: Arrow Laws

The Arrow law (ab)c=a(bc),(a \ggg b)\ggg c=a \ggg (b \ggg c), represents associativity of composition.

Diagram 6
Figure 6: Arrow Laws

The Arrow law first(ab)=first(a)first(b)\text{first}(a\ggg b)=\text{first}(a)\ggg \text{first}(b) encodes functoriality of ×W:CC- \times W: C \to C.

Diagram 7
Figure 7: Arrow Laws

The Arrow law first(a)arr(π 1)=arr(π 1)a\text{first}(a)\ggg \text{arr}(\pi_{1})=\text{arr}(\pi_{1})\ggg a express naturality of the counit ×Wid C- \times W \to \text{id}_{C}, i.e., the first projection maps.

Diagram 8
Figure 8: Arrow Laws

The Arrow law first(a)arr(α)=arr(α)first(first(a))\text{first}(a)\ggg \text{arr}(\alpha)=\text{arr}(\alpha)\ggg \text{first}(\text{first}(a)) asks that first\text{first} play nicely with associators.

Diagram 9
Figure 9: Arrow Laws

The Arrow law first(a)arr(id×f)=arr(id×f)first(a)\text{first}(a)\ggg \text{arr}(\text{id} \times f)=\text{arr}(\text{id} \times f)\ggg \text{first}(a) is an interchange law which says id×g:(×W)(×W)\text{id} \times g:(- \times W) \to (- \times W') is a natural transformation for every g:WWg:W \to W' in CC.

Diagram 10
Figure 10: Arrow Laws

Two Arrow laws trivialise as a result of our example, so diagrams aren’t produced. The first such law is arr(f;g)=arr(f)arr(g).\text{arr}(f;g)=\text{arr}(f)\ggg \text{arr}(g). For our example, this law trivialises, as :=composition\ggg : = \text{composition} and arr:=id (Y X).\text{arr} := \text{id}_{(Y^X)}. The second law to trivialise is first(arr(f))=arr(f×id)\text{first}(\text{arr}(f))=\text{arr}(f \times \text{id}) since we have set first(f):=f×id.\text{first}(f) := f \times \text{id}.

Freyd Categories

To understand Freyd categories, we must first define what a symmetric premonoidal category is.

Definition. A symmetric premonoidal category includes:

  • An object II (unit).
  • Natural transformations that define how objects interact:
    • Associativity: α:(xy)zx(yz)\alpha : (x \otimes y) \otimes z \to x \otimes (y \otimes z)
    • Left unitor: λ:xIx\lambda : x \otimes I \to x
    • Right unitor: ρ:Ixx\rho : I \otimes x \to x
    • Symmetry: σ:xyyx\sigma : x \otimes y \to y \otimes x
  • All components are central .

A morphism f:xxf : x \to x' is central if g:yy,fy;xg=xg;fy\forall g:y \to y', \quad f \otimes y ; x' \otimes g = x \otimes g ; f \otimes y'

Now, we can define a Freyd category, recalling the definition from the introduction.

Definition. A Freyd category consists of:

  • A category CC with finite products.
  • A symmetric premonoidal category KK.
  • An identity-on-objects functor J:CKJ : C \to K that:
    • Preserves symmetric premonoidal structure.
    • Ensures J(f)J(f) is always central.

Arrows vs Freyd Categories: Similarities and Differences

At first glance, the definition of a Freyd category appears strikingly similar to that of an Arrow. This apparent similarity led to the folklore belief that they were equivalent structures.

A Freyd category consists of two categories CC and KK with an identity-on-objects functor J:CKJ: C \to K, where: - CC has finite products - KK is symmetric premonoidal (with a functor z&#8722; \otimes z) - JJ maps finite products in CC to the premonoidal structure in KK

In our culinary metaphor, this loosely translates to: - CC: The idealized recipes (Haskell types and functions) - KK: The real-world kitchen operations (computations represented by the Arrow type Ar(x,y)\text{Ar}(x,y)) - JJ: The translation process (via arr, embedding pure functions) - Composition in KK: The sequencing of operations (via >>>) - Premonoidal structure in KK: The ability to process pairs (via first)

Recalling the how we’ve interpreted Arrows in the cullinary setting, the apparent correspondence between Arrows and Fryed categories seemes quite natural. In fact, for many years the two concepts were thought to be two ways of speaking about the same thing among those in the programming languages community.

However, Atkey’s work revealed a crucial distinction: Arrows are more general than Freyd categories . The key difference lies in how they handle inputs:

  • Freyd categories allow only a single input to computations
  • Arrows support two separate inputs:
    • One may be structured (modeled using comonads)
    • This additional flexibility allows Arrows to represent computations that Freyd categories cannot

To bridge this gap, Atkey introduced the concept of indexed Freyd categories , which can model two structured inputs. The relationship can be summarized as: Arrows are equivalent to Closed Indexed Freyd Categories.

In our culinary metaphor, we can understand this relationship as follows: a Freyd category is like a restaurant that can only take one order at a time (a single input), while Arrows are like a more sophisticated establishment that can handle both individual orders and special requests that come with their own context (two inputs, one potentially structured). The closed indexed Freyd categories that Atkey identifies represent the perfect middle ground — restaurants that can efficiently manage multiple orders with specialized instructions while maintaining the core operational principles that make kitchens function. This is particularly valuable when preparing complex “quantum dishes” where ingredients might be entangled and interact with each other in non-local ways.

Diagram 11
Figure 11: Arrows vs. Freyd Categories. Arrows support two inputs (one potentially structured) and are equivalent to Closed Indexed Freyd Categories, which generalize standard Freyd Categories that handle only single inputs.

This distinction helps explain why Arrows have proven particularly useful in domains like quantum computing, where managing multiple inputs with complex relationships is essential.

R. Atkey’s paper finds the relationship between Arrows and different constraints on Freyd categories as follows:

Diagram 12
Figure 12: Relationship Between Structures


Key Insights: - Arrows can be defined both as monoids in categories of strong profunctors and operationally through concrete morphisms (arr\text{arr}, \ggg, first\text{first}) - Freyd categories formalize the relationship between pure functions and effectful computations using symmetric premonoidal structure - Despite the folklore belief, Arrows are strictly more general than Freyd categories because they can handle two separate inputs (one potentially structured) - Arrows are equivalent to closed indexed Freyd categories, bridging the conceptual gap


Applications and Questions The main goal of our Adjoint School project was to structure effects in quantum programming languages using generalizations of monads. Relative monads are a popular generalization of monads. These monads need not be endofunctors, and they’re known to generalize Arrows as well. Since we already know how to structure quantum effects using Arrows, it follows that it should be theoretically possible to structure quantum effects using relative monads.

Arrows’ capacity to handle multiple inputs with a single, potentially structured output offers tractability that is particularly useful in quantum computing. Particles in quantum systems can be in entangled states, where the manipulation of one particle influences others in real time, irrespective of distance. This non-local interaction can be modeled through Arrows’ ability to combine several inputs while keeping track of their interrelationships.

Our group investigated the possibility of doing exactly this. The main technical issue arises from the fact that the way Arrows have been implemented in Haskell to structure quantum effects does not provide a categorical semantics for the problem.

For our ACT2025 presentation, we were able to construct a relative monad capable of handling classical control in the quantum setting, but the following questions still remain:

  • Can one build a relative monad to model quantum effects?

  • If so, how might an implementation of these ideas in Haskell compare to Arrow-based approaches?

The ride from burrito monads to Arrow kitchens has carried us farther than we anticipated, illustrating that even established mathematical folklore sometimes requires precise re-evaluation. As we continue to learn about these structures, we hope this post will motivate others to participate in the exploration of these tools and their use in quantum computing and beyond.

September 12, 2025

Matt von HippelWhat You’re Actually Scared of in Impostor Syndrome

Academics tend to face a lot of impostor syndrome. Something about a job with no clear criteria for success, where you could always in principle do better and you mostly only see the cleaned-up, idealized version of others’ work, is a recipe for driving people utterly insane with fear.

The way most of us talk about that fear, it can seem like a cognitive bias, like a failure of epistemology. “Competent people think they’re less competent than they are,” the less-discussed half of the Dunning-Kruger effect.

(I’ve talked about it that way before. And, in an impostor-syndrome-inducing turn of events, I got quoted in a news piece in Nature about it.)

There’s something missing in that perspective, though. It doesn’t really get across how impostor syndrome feels. There’s something very raw about it, something that feels much more personal and urgent than an ordinary biased self-assessment.

To get at the core of it, let me ask a question: what happens to impostors?

The simple answer, the part everyone will admit to, is to say they stop getting grants, or stop getting jobs. Someone figures out they can’t do what they claim, and stops choosing them to receive limited resources. Pretty much anyone with impostor syndrome will say that they fear this: the moment that they reach too far, and the world decides they aren’t worth the money after all.

In practice, it’s not even clear that that happens. You might have people in your field who are actually thought of as impostors, on some level. People who get snarked about behind their back, people where everyone rolls their eyes when they ask a question at a conference and the question just never ends. People who are thought of as shiny storytellers without substance, who spin a tale for journalists but aren’t accomplishing anything of note. Those people…aren’t facing consequences at all, really! They keep getting the grants, they keep finding the jobs, and the ranks of people leaving for industry are instead mostly filled with those you respect.

Instead, I think what we fear when we feel impostor syndrome isn’t the obvious consequence, or even the real consequence, but something more primal. Primatologists and psychologists talk about our social brain, and the role of ostracism. They talk about baboons who piss off the alpha and get beat up and cast out of the group, how a social animal on their own risks starvation and becomes easy prey for bigger predators.

I think when we wake up in a cold sweat remembering how we had no idea what that talk was about, and were too afraid to ask, it’s a fear on that level that’s echoing around in our heads. That the grinding jags of adrenaline, the run-away-and-hide feeling of never being good enough, the desperate unsteadiness of trying to sound competent when you’re sure that you’re not and will get discovered at any moment…that’s not based on any realistic fears about what would happen if you got caught. That’s your monkey-brain, telling you a story drilled down deep by evolution.

Does that help? I’m not sure. If you manage to tell your inner monkey that it won’t get eaten by a lion if its friends stop liking it, let me know!

Scott Aaronson Quantum Information Supremacy

I’m thrilled that our paper entitled Demonstrating an unconditional separation between quantum and classical information resources, based on a collaboration between UT Austin and Quantinuum, is finally up on the arXiv. I’m equally thrilled that my coauthor and former PhD student William Kretschmer — who led the theory for this project, and even wrote much of the code — is now my faculty colleague at UT Austin! My physics colleague Nick Hunter-Jones and my current PhD student Sabee Grewal made important contributions as well. I’d especially like to thank the team at Quantinuum for recognizing a unique opportunity to test and showcase their cutting-edge hardware, and collaborating with us wild-eyed theorists to make it happen. This is something that, crucially, would not have been feasible with the quantum computing hardware of only a couple years ago.

Here’s our abstract, which I think explains what we did clearly enough, although do read the paper for more:

A longstanding goal in quantum information science is to demonstrate quantum computations that cannot be feasibly reproduced on a classical computer. Such demonstrations mark major milestones: they showcase fine control over quantum systems and are prerequisites for useful quantum computation. To date, quantum advantage has been demonstrated, for example, through violations of Bell inequalities and sampling-based quantum supremacy experiments. However, both forms of advantage come with important caveats: Bell tests are not computationally difficult tasks, and the classical hardness of sampling experiments relies on unproven complexity-theoretic assumptions. Here we demonstrate an unconditional quantum advantage in information resources required for a computational task, realized on Quantinuum’s H1-1 trapped-ion quantum computer operating at a median two-qubit partial-entangler fidelity of 99.941(7)%. We construct a task for which the most space-efficient classical algorithm provably requires between 62 and 382 bits of memory, and solve it using only 12 qubits. Our result provides the most direct evidence yet that currently existing quantum processors can generate and manipulate entangled states of sufficient complexity to access the exponentiality of Hilbert space. This form of quantum advantage — which we call quantum information supremacy — represents a new benchmark in quantum computing, one that does not rely on unproven conjectures.

I’m very happy to field questions about this paper in the comments section.


Unrelated Announcement: As some of you might have seen, yesterday’s Wall Street Journal carried a piece by Dan Kagan-Kans on “The Rise of ‘Conspiracy Physics.'” I talked to the author for the piece, and he quoted this blog in the following passage:

This resentment of scientific authority figures is the major attraction of what might be called “conspiracy physics.” Most fringe theories are too arcane for listeners to understand, but anyone can grasp the idea that academic physics is just one more corrupt and self-serving establishment. The German physicist Sabine Hossenfelder has attracted 1.72 million YouTube subscribers in part by attacking her colleagues: “Your problem is that you’re lying to the people who pay you,” she declared. “Your problem is that you’re cowards without a shred of scientific integrity.”

In this corner of the internet, the scientist Scott Aaronson has written, “Anyone perceived as the ‘mainstream establishment’ faces a near-insurmountable burden of proof, while anyone perceived as ‘renegade’ wins by default if they identify any hole whatsoever in mainstream understanding.”

September 11, 2025

Doug NatelsonDOE Experimental Condensed Matter Physics PI Meeting 2025 - Day 3 and wrap-up

 A few more interesting tidbits from the concluding half-day of the DOE ECMP PI meeting:

Unfortunately I missed the last talk because of the need to head to the airport.  Overall, the meeting was very good.  Program PI meetings can tend to become less about telling coherent scientific stories and more about trying to show everything someone has done in the last three years.  This meeting avoided that, with clear talks that generally focused on one main result, and that made it much more engaging.  As good as tools for virtual gatherings have become, there really is no substitute for an in-person event when you can just talk to someone by the coffee about some new idea.

John BaezCategories for Public Health Modeling

How, exactly, can category theory help modeling in public health? I wrote a paper about this with two people who helped run Canada’s COVID modeling, together with a software engineer and a mathematician at the Topos Institute:

• John Baez, Xiaoyan Li, Sophie Libkind, Nathaniel D. Osgood and Eric Redekopp, A categorical framework for modeling with stock and flow diagrams, in Mathematics of Public Health: Mathematical Modelling from the Next Generation, eds. Jummy David and Jianhong Wu, Springer, 2023, pp. 175-207.

Anything you can with category theory, you can also do without it—just like you can cross the Alps without shoes. But categorical methods make public health modeling easier in a lot of ways.

The introduction lists a few of these ways. Then the paper goes on to provide details, including a long appendix showing actual Julia code for our software.

But here’s the basic idea:

Many people working on epidemiological modeling like diagrams because they provide easily understandable but informal steps towards a mathematically rigorous formulation of a model in terms of ordinary differential equations (ODEs). But ODEs are typically opaque to non-modelers—including the interdisciplinary members of the teams that typically are required for impactful models.

The tradition of modeling called System Dynamics places a premium on engagement with stakeholders, so it offers a modeling approach centered around diagrams. This approach commonly proceeds in a manner that depicts model structure using successively more detailed models.

The process starts with a ‘causal loop diagram’ illustrating causal connections and feedback loops:

It then often proceeds to a ‘system structure diagram’, which distinguishes stocks from flows but still lacks quantitative information. The next step is to construct a stock and flow diagram’:

This diagram is visually identical to the system structure diagram, but it also includes formulae (at the bottom here).

The stock–flow diagram is treated as the durable end result of this modeling process, since it uniquely specifies a system of first-order ODEs. System Dynamics modeling typically then alternates between assessing scenario outcomes resulting from numerically integrating the ODEs, performing other analyses (e.g., identifying location or stability of equilibria), and elaborating the stock-flow diagram.

While each of the 3 types of diagrams in the System Dynamics tradition is recommended by visual accessibility, the traditional approach suffers from a number of practical shortcomings:

Monolithic models: Models are traditionally built up in a monolithic fashion, leading ultimately to a single large piece of code. Drawn as a single diagram, a model can be extremely complex. For example, here is Canada’s main model of COVID during the pandemic, put together by Nathaniel Osgood and Xiaoyan Li, made using the commercially available software called AnyLogic:


Click to enlarge. If it looks like a huge mess, that’s part of the point.

Working with a single huge model like this inhibits independent simultaneous work by multiple modelers. Lack of model modularity further prevents effective reuse of particular model elements. If elements of other models are used, they are commonly copy-and-pasted into the developing model, with the source and destination then evolving independently. Such separation can lead to a proliferation of conceptually overlapping models in which a single conceptual change requires corresponding updates in several successive models.

The curse of dimensionality: Modelers refine simple models by ‘stratifying’ them, subdividing stocks into smaller stocks. For example, the ‘infected’ stock might be stratified into ‘infected male’ and ‘infected female’. While stratification is a key tool for representing heterogeneity, stratification commonly requires modifications across the breadth of a model—stocks, flows, derived quantities, and many parameters. When stratification involves multiple dimensions of heterogeneity, it can lead to a proliferation of terms in the ODEs. For example, rendering a model characterizing both COVID-19 into a model also characterizes influenza would require that each COVID-19 state to be replicated for each stage in the natural history of influenza. Represented visually, this stratification leads to a multi-dimensional lattice, commonly with progression proceeding along several dimensions of the lattice. Because of the unwieldy character of the diagram, the structure of the model is obscured. Adding, removing, or otherwise changing dimensions of heterogeneity routinely leads to pervasive changes across the model.

Privileging ODE semantics: The structure of causal loop diagrams, system structure diagrams and stock-flow diagrams characterizes general state and accumulations, transitions and posited causal relations—including induced feedbacks—amongst variables. Nothing about such a characterization restricts its meaning to ordinary differential equations; indeed, many other interpretations and uses of these diagrams are possible. However, existing software privileges an ODE interpretation for stock-flow diagrams, while sometimes allowing for secondary analyses in ad hoc way—for example, identifying causal loops associated with the model, or verifying dimensional homogeneity in dimensionally annotated models. Conducting other sorts of analyses—such as computation of eigenvalue elasticities or loop gains, analysis as a stochastic transition system, or other methods such as particle filtering, particle Markov chain Monte Carlo, or Kalman filtering—typically requires bespoke software for reading, representing and analyzing stock-flow models.

Divergence of model representations: Although the evolution from causal loop diagrams to system structure diagrams to stock-flow models is one of successive elaboration and informational enrichment, existing representations treat these as entirely separate characterizations and fail to capture the logical relationships between them. Such fragmentation commonly induces inconsistent evolution. Indeed, in many projects, the evolution of stock-flow diagrams renders the earlier, more abstract formulations obsolete, and the focus henceforth rests on the stock-flow diagrams.

What is less widely appreciated is that beyond their visual transparency and capacity to be lent a clear ODE semantics, the 3 kinds of diagrams I mentioned each have a precise mathematical structure—a corresponding grammar, as it were. This algebraic structure, called the ‘syntax’ of tehse diagrams, can be characterized using category theory. Formalizing the syntax this way lends precise meaning to the process of ‘composing’ models (building them out of smaller parts), stratifying them, and other operations. Explicitly characterizing the syntax in software also allows for diagrams to be represented, manipulated, composed, transformed, and flexibly analyzed in software that implements the underlying mathematics.

Formalizing the mathematics of diagram-based models using category theory and capturing it in software offers many benefits. Our paper discusses and demonstrates just a few:

Separation of syntax and semantics. Category theory gives tools to separate the formal structure, or ‘syntax’, of diagram-based models from the uses to which they are put, or ‘semantics’. The syntax lives in one category, which can then be mapped to various different semantic categories using various functors. This separation permits great flexibility in applying different semantics to the same model. With appropriate software design, this decoupling can allow the same software to support a diverse array of analyses, which can be supplemented over time.

Reuse of structure. Category theory provides a structured way to build complex diagrams by composing small reusable pieces. Diagrams are morphisms in a monoidal category, and you build bigger diagrams by composing and tensoring these morphisms. With software support, modeling frameworks can allow for saving models and retrieving them for reuse as parts of many different models. For example, in public health a diagram representing contact tracing can be reused across diagrams addressing different pathogens.

Modular stratification. A categorical foundation further supports a structured method to build stratified diagrams out of modular, reusable, largely orthogonal pieces. This method is called taking a ‘pullback’. In contrast to the global changes commonly required to a diagram and the curse of dimensionality that traditionally arises when stratifying a diagram, categorically-founded stratification methods allow for crisply characterizing a stratified diagram as built from simpler diagrams, one for each heterogeneity or progression dimension.

Our paper goes into detail about how all this works. Elsewhere we have longer lists of what’s bad about current modeling practice, and how we hope to improve it. But I hope this helps a bit.

Doug NatelsonDOE Experimental Condensed Matter Physics PI Meeting 2025 - Day 2

It was another very full day.   I had to pop in and out to attend to some things so I didn't get everything, but here are some physics items I learned:

  • Dillon Fong introduced me to a technique I didn't know about before, x-ray photon correlation spectroscopy (see this paper).  You can look at time correlations of x-ray speckle near a particular Bragg spot and learn about dynamics and kinetics of transitions and materials growth.  Very cute.
  • Charles Ahn presented work on high magnetic field superconductivity in Nd(1-x)Eu(x)NiO2, and I learned about the Jaccarino-Peter effect, in which an external magnetic field can counter the interaction between magnetic dopants and the conduction electrons.  This leads to "reentrant" superconductivity at high magnetic fields. 
  • Danny Phelan showed that you can have two different crystal structures for La3Ni2O7, one that is stacked bilayers ("2222"), and one that is stacked monolayer/trilayer ("1313").  
  • Ian Fisher talked about using the elastocaloric effect (rapidly and therefore adiabatically stretch or compress a material, leading to a change in its temperature) to identify phase transitions, since the effect is proportional to \( (\partial S/\partial \epsilon)_{T}\), the change in entropy with strain.
  • Dan Dessau presented an interesting analysis of data in cuprates suggesting a form for the electronic self-energy that is called a power law liquid, and that this analysis implies that there is not a quantum critical point under the middle of the superconducting dome.
  • Jak Chakhalian showed that epitaxially growing an iridate Weyl semimetal directly on top of insulating Dy2Ti2O7 spin ice leads to a dramatic anisotropic magnetoresistance at high in-plane fields that identifies interesting previously unknown physics.
  • Daniel Rhodes showed some pretty work on superconductivity in T_d-MoTe2.  This material is extremely air-sensitive, and all of the device fabrication has to be done with great care in a glovebox.  This led to the following exchange.  Audience question: "It is notoriously difficult to make electrical contact to this material.  How did you do this?"  Answer: "Through tears and blood."  This was followed by a serious answer that concluded "The glovebox is always the problem."

September 10, 2025

John PreskillNicole’s guide to writing research statements

Sunflowers are blooming, stores are trumpeting back-to-school sales, and professors are scrambling to chart out the courses they planned to develop in July. If you’re applying for an academic job this fall, now is the time to get your application ducks in a row. Seeking a postdoctoral or faculty position? Your applications will center on research statements. Often, a research statement describes your accomplishments and sketches your research plans. What do evaluators look for in such documents? Here’s my advice, which targets postdoctoral fellowships and faculty positions, especially for theoretical physicists.

  • Keep your audience in mind. Will a quantum information theorist, a quantum scientist, a general physicist, a general scientist, or a general academic evaluate your statement? What do they care about? What technical language do and don’t they understand?
  • What thread unites all the projects you’ve undertaken? Don’t walk through your research history chronologically, stepping from project to project. Cast the key projects in the form of a story—a research program. What vision underlies the program?
  • Here’s what I want to see when I read a description of a completed project.
    • The motivation for the project: This point ensures that the reader will care enough to read the rest of the description.
    • Crucial background information
    • The physical setup
    • A statement of the problem
    • Why the problem is difficult or, if relevant, how long the problem has remained open
    • Which mathematical toolkit you used to solve the problem or which conceptual insight unlocked the solution
    • Which technical or conceptual contribution you provided
    • Whom you collaborated with: Wide collaboration can signal a researcher’s maturity. If you collaborated with researchers at other institutions, name the institutions and, if relevant, their home countries. If you led the project, tell me that, too. If you collaborated with a well-known researcher, mentioning their name might help the reader situate your work within the research landscape they know. But avoid name-dropping, which lacks such a pedagogical purpose and which can come across as crude.
    • Your result’s significance/upshot/applications/impact: Has a lab based an experiment on your theoretical proposal? Does your simulation method outperform its competitors by X% in runtime? Has your mathematical toolkit found applications in three subfields of quantum physics? Consider mentioning whether a competitive conference or journal has accepted your results: QIP, STOC, Physical Review Letters, Nature Physics, etc. But such references shouldn’t serve as a crutch in conveying your results’ significance. You’ll impress me most by dazzling me with your physics; name-dropping venues instead can convey arrogance.
  • Not all past projects deserve the same amount of space. Tell a cohesive story. For example, you might detail one project, then synopsize two follow-up projects in two sentences.
  • A research statement must be high-level, because you don’t have space to provide details. Use mostly prose; and communicate intuition, including with simple examples. But sprinkle in math, such as notation that encapsulates a phrase in one concise symbol.

  • Be concrete, and illustrate with examples. Many physicists—especially theorists—lean toward general, abstract statements. The more general a statement is, we reason, the more systems it describes, so the more powerful it is. But humans can’t visualize and intuit about abstractions easily. Imagine a reader who has four minutes to digest your research statement before proceeding to the next 50 applications. As that reader flys through your writing, vague statements won’t leave much of an impression. So draw, in words, a picture that readers can visualize. For instance, don’t describe only systems, subsystems, and control; invoke atoms, cavities, and lasers. After hooking your reader with an image, you can generalize from it.
  • A research statement not only describes past projects, but also sketches research plans. Since research covers terra incognita, though, plans might sound impossible. How can you predict the unknown—especially the next five years of the unknown (as required if you’re applying for a faculty position), especially if you’re a theorist? Show that you’ve developed a map and a compass. Sketch the large-scale steps that you anticipate taking. Which mathematical toolkits will you leverage? What major challenge do you anticipate, and how do you hope to overcome it? Let me know if you’ve undertaken preliminary studies. Do numerical experiments support a theorem you conjecture?
  • When I was applying for faculty positions, a mentor told me the following: many a faculty member can identify a result (or constellation of results) that secured them an offer, as well as a result that earned them tenure. Help faculty-hiring committees identify the offer result and the tenure result.
  • Introduce notation before using it. If you use notation and introduce it afterward, the reader will encounter the notation; stop to puzzle over it; tentatively continue; read the introduction of the notation; return to the earlier use of the notation, to understand it; and then continue forward, including by rereading the introduction of the notation. This back-and-forth breaks up the reading process, which should flow smoothly.
  • Avoid verbs that fail to relate that you accomplished anything: “studied,” “investigated,” “worked on,” etc. What did you prove, show, demonstrate, solve, calculate, compute, etc.?

  • Tailor a version of your research statement to every position. Is Fellowship Committee X seeking biophysicists, statistical physicists, mathematical physicists, or interdisciplinary scientists? Also, respect every application’s guidelines about length.
  • If you have room, end the statement with a recap and a statement of significance. Yes, you’ll be repeating ideas mentioned earlier. But your reader’s takeaway hinges on the last text they read. End on a strong note, presenting a coherent vision.

  • Writing is rewriting, a saying goes. Draft your research statement early, solicit feedback from a couple of mentors, edit the draft, and solicit more feedback.

Doug NatelsonDOE Experimental Condensed Matter Physics PI Meeting 2025 - Day 1

That was a full day.  Here are some things I learned, beyond the fact that the ballroom here is clearly kept at about 15°C by default.  (Apologies for not getting everything....)

  • About 40% of the DOE ECMP program is related to 2D materials these days.
  • Long Ju showed some interesting work trying to understand rhombohedral (ABC-stacked) 5-layer graphene encapsulated by hBN.  Trying to get rid of moiré effects from the hBN/graphene interfaces leads not to more robust quantum anomalous Hall response, but instead leads to very peculiar superconductivity that survives up to very large in-plane and moderately large out-of-plane magnetic fields. This happens in the same regime of charge and gate that would otherwise show QAH.  Looks like some kind of chiral superconductivity that may be topological.
  • Andrea Young, meanwhile, in fewer layer rhombohedral systems, showed experiments pointing to superconductivity happening at the verge of a canting transition, where spins are reorienting.
  • Eva Andrei gave a nice talk looking at the variety of states one can get when interfacing moiré systems with other moiré systems, and explaining what is meant by intercrystals.  
  • Gleb Finkelstein showed how a measurement intended to look at shot noise instead became a very cute noise thermometry probe of thermal transport at the boundary between (graphene) quantum Hall currents and a superconducting electrode.
  • Xiao-Xiao Zhang showed a really cute experiment, where the resonance of a drumhead made from an atomically thin film of MnPS3 convey information about magnetic transitions in that material as a function of magnetic field.
  • Dan Ralph gave a nice talk about the challenges of electrically generating currents of properly oriented spins to drive magnetic switching in films magnetized perpendicular to the film plane, for spin-orbit torque memories (and fundamental understanding).
  • Philip Kim gave a great overview of some remarkable results in electronic interferometers made on graphene, in which telegraph noise shows signatures
  • Lu Li spoke about recent measurements showing magnetic oscillations and specific heat signatures of possible neutral fermions in a kagome lattice Mott insulator.
  • Xavier Roy talked about CeSiI, a 2D material that is also a heavy fermion metal.  This and its related compounds look like a fascinating family of (unfortunately extremely air sensitive) materials. 
  • Harold Hwang gave a great overview of recent work in nickelate superconductors, highlighting the similarities to the cuprates as well as the profound differences (like how electronic configurations other than d9 can also lead to superconductivity).

September 09, 2025

Doug NatelsonDOE experimental condensed matter PI meeting, + other items

This week I am attending the every-two-years DOE Experimental Condensed Matter Physics PI meeting.  Previously I have written up highlights of these meetings (see here, here, here, here, here), though two years I was unable to do so because I was attending virtually.  I will do my best to hit some high points (though I will restrict myself to talking only about already published work, to avoid any issues of confidentiality).  

In the meantime, here are a couple of topics of interest from the last couple of weeks.  

  • I just learned about the existence of Mathos AI, an AI product that can function as a math solver and calculator, as well as a tutor for students.  It is pretty impressive.
  • I liked this historical piece about Subrahmanyan Chandrasekhar (he of the “Chandrasekhar limit”, which describes the degeneracy physics + gravitation that limits the upper size of compact stellar objects like white dwarfs and neutron stars before they collapse into black holes) and his interactions with Stephen Hawking.  It's pretty humanizing to see an intellectual giant like Chandra sending a brief letter to Hawking in 1967 asking for advice on what to read so that Chandra can understand Hawking’s work on singularities in cosmology.  Hawking’s handwritten response is clear and direct.
  • In an online discussion about what people will do if Google decides to stop supporting Google Scholar, I was introduced to OpenAlex.  This seems like an interesting, also-free alternative.  Certainly worth watching.  There is no obvious reason to think that Google Scholar is going away, but Alphabet has retired many free products, and it’s far from obvious how they are making any money on this.  Anyone from Google who reads the blog, please chime in.  (Note to self:  keep regularly backing up this blog, since blogger is also not guaranteed future existence.)


September 08, 2025

John BaezIs Category Theory Being Co-opted?

Is category theory being co-opted? These authors think so:

• Esteban Montero and Brandon Baylor, Category theory is being co-opted, Holon Substack, 7 September 2025.

Category theory offers a powerful new way of viewing the world… if we embrace its lessons and are willing to think in brand new ways. But what are the chances of that? It could easily become just another way for the rich and powerful to get more rich and powerful while the world burns. Applied category theory is already heavily funded by the military and tech firms.

Luckily, the UK has injected money into category theory for developing safeguards to AI. It’s risky, but it might do some good. And some of us are using category theory to develop new software for epidemiology, and soon labor economics, trying to model aspects of the green transition.

Montero and Baylor’s article is interesting—but it ignores those initiatives. It focuses heavily on the supposed ills of set theory—but I don’t think escaping set theory is either necessary or sufficient to free up our ways of thinking.

I have plenty of other complaints about this essay:

• Much as I like applied category theory, I don’t think our work so far “demonstrates the immense practical power of categorical thinking.” I think the jury is still out, which is why I keep working hard on this.

• The focus on Buddhism makes me uneasy. Buddhism has some good ideas, but I don’t think the cause of good new mathematics should be shackled to this religion.

• It seems like a bad idea, tactically, to combine this sort of essay with a plea for funding—even if you really want funding!

Nonetheless, not enough people are talking about the remarkable transformation of category theory into an ever-more-practical discipline, and the question of whether it will be truly revolutionary, or just another tool.

So until someone writes a better essay about this, it’s worth reading this one!

I particularly like their vision of how categorical network theory could be part of a new civilization:

Systems Engineering: For a systems engineer designing a smart grid, a logistics network, or a complex software architecture, the most important thing is not the individual components, but the rules of their interaction. Category theory provides a formal language for designing systems not by specifying parts, but by defining relationships, interfaces, and compositions.

Ecological Thinking: In ecology, an organism cannot be understood apart from its niche—the web of relationships it has with its environment and other organisms. The categorical perspective aligns with this systemic view, seeing entities as defined by their context.

A New Civilization: At the largest scale, this relational philosophy suggests a different way of organizing society. Instead of a collection of atomized individuals, it envisions a system of interconnected stakeholders, where the health of the whole is determined by the quality of the relationships between the parts.

We just need to work out a few details.

September 05, 2025

Matt von HippelThe Rocks in the Ground Era of Fundamental Physics

It’s no secret that the early twentieth century was a great time to make progress in fundamental physics. On one level, it was an era when huge swaths of our understanding of the world were being rewritten, with relativity and quantum mechanics just being explored. It was a time when a bright student could guide the emergence of whole new branches of scholarship, and recently discovered physical laws could influence world events on a massive scale.

Put that way, it sounds like it was a time of low-hanging fruit, the early days of a field when great strides can be made before the easy problems are all solved and only the hard ones are left. And that’s part of it, certainly: the fields sprung from that era have gotten more complex and challenging over time, requiring more specialized knowledge to make any kind of progress. But there is also a physical reason why physicists had such an enormous impact back then.

The early twentieth century was the last time that you could dig up a rock out of the ground, do some chemistry, and end up with a discovery about the fundamental laws of physics.

When scientists like Curie and Becquerel were working with uranium, they didn’t yet understand the nature of atoms. The distinctions between elements were described in qualitative terms, but only just beginning to be physically understood. That meant that a weird object in nature, “a weird rock”, could do quite a lot of interesting things.

And once you find a rock that does something physically unexpected, you can scale up. From the chemistry experiments of a single scientist’s lab, countries can build industrial processes to multiply the effect. Nuclear power and the bomb were such radical changes because they represented the end effect of understanding the nature of atoms, and atoms are something people could build factories to manipulate.

Scientists went on to push that understanding further. They wanted to know what the smallest pieces of matter were composed of, to learn the laws behind the most fundamental laws they knew. And with relativity and quantum mechanics, they could begin to do so systematically.

US particle physics has a nice bit of branding. They talk about three frontiers: the Energy Frontier, the Intensity Frontier, and the Cosmic Frontier.

Some things we can’t yet test in physics are gated by energy. If we haven’t discovered a particle, it may be because it’s unstable, decaying quickly into lighter particles so we can’t observe it in everyday life. If these particles interact appreciably with particles of everyday matter like protons and electrons, then we can try to make them in particle colliders. These end up creating pretty much everything up to a certain mass, due to a combination of the tendency in quantum mechanics for everything that can happen to happen, and relativity’s E=mc^2. In the mid-20th century these particle colliders were serious pieces of machinery, but still small enough to make industrial: now, there are so-called medical accelerators in many hospitals based on their designs. But current particle accelerators are a different beast, massive facilities built by international collaborations. This is the Energy Frontier.

Some things in physics are gated by how rare they are. Some particles interact only very faintly with other particles, so to detect them, physicists have to scan a huge chunk of matter, a giant tank of argon or a kilometer of antarctic ice, looking for deviations from the norm. Over time, these experiments have gotten bigger, looking for more and more subtle effects. A few weird ones still fit on tabletops, but only because they have the tools to measure incredibly small variations. Most are gigantic. This is the Intensity Frontier.

Finally, the Cosmic Frontier looks for the unknown behind both kinds of gates, using the wider universe to look at events with extremely high energy or size.

Pushing these frontiers has meant cleaning up our understanding of the fundamental laws of physics up to these frontiers. It means that whatever is still hiding, it either requires huge amounts of energy to produce, or is an extremely rare, subtle effect.

That means that you shouldn’t expect another nuclear bomb out of fundamental physics. Physics experiments are already working on vast scales, to the extent that a secret government project would have to be smaller than publicly known experiments, in physical size, energy use, and budget. And you shouldn’t expect another nuclear power plant, either: we’ve long passed the kinds of things you could devise a clever industrial process to take advantage of at scale.

Instead, new fundamental physics will only be directly useful once we’re the kind of civilization that operates on a much greater scale than we do today. That means larger than the solar system: there wouldn’t be much advantage, at this point, of putting a particle physics experiment on the edge of the Sun. It means the kind of civilization that tosses galaxies around.

It means that right now, you won’t see militaries or companies pushing the frontiers of fundamental physics, unlike the way they might have wanted to at the dawn of the twentieth century. By the time fundamental physics is useful in that way, all of these actors will likely be radically different: companies, governments, and in all likelihood human beings themselves. Instead, supporting fundamental physics right now is an act of philanthropy, maintaining a practice because it maintains good habits of thought and produces powerful ideas, the same reasons organizations support mathematics or poetry. That’s not nothing, and fundamental physics is still often affordable as philanthropy goes. But it’s not changing the world, not the way physicists did in the early twentieth century.

Jordan EllenbergYu Darvish has beaten 29 out of the 30 MLB teams

But not the Orioles, against whom he is now 0-3. He got beat by Dylan Bundy in the 2016 game, then by Grayson Rodriguez (remember Grayson Rodriguez?) in 2023, and by Tyler Wells this week. That’s only the regular season, though. The game I really remember is the 2012 AL wild card game, the first playoff game for the Orioles in more than a decade. Darvish, then an MLB rookie, started for the Rangers. And we beat him then, too. I just claimed to really remember that game, but if you asked me who won it for the Orioles? You could give me fifty guesses and I wouldn’t have come up with Joe Saunders.

I would have thought it was hard to beat every team, but in the balanced schedule / interleague schedule it’s gotten a lot easier. Of course Wikipedia has a page of everybody who’s done it. Our old friend Kevin Gausman has managed to beat all 30 teams despite having only 110 wins. Maybe even more impressive is Jameson Taillon (who I saw beat the Brewers a couple of weeks ago in my first-ever trip to Wrigley Field, a game I never got around to blogging about), who is 0-1 against the Dodgers in 4 tries, but who has beaten all 29 other teams with only 80 career wins!

Wikipedia is really an incredible triumph of human cooperation. Why does it work so well? How is it so unpolluted? How is it that when some very weird niche question like “which pitchers have beaten all 30 teams,” comes to my mind, some human being has already compiled this and put it there? I don’t know. But I’m glad it exists.

I don’t know why there are so many baseball posts right now. They’re faster to write than math posts. More math posts to come, I promise! (But there’s also one more Orioles post I want to write…)

Update: This is not the Orioles post I have planned, but I should point out that perhaps even more impressively than any of the above feats of win distribution is that the 2025 Orioles, with only 64 wins to their name so far, have already beaten 28 of the 29 other teams in baseball, getting season-swept only by the Minnesota Twins.

September 04, 2025

Scott Aaronson For the record

In response to my recent blog posts, which expressed views that are entirely boring and middle-of-the-road for Americans as a whole, American Jews, and Israelis (“yes, war to destroy Hamas is basically morally justified, even if there are innocent casualties, as the only possible way to a future of coexistence and peace”)—many people declared that I was a raving genocidal maniac who wants to see all Palestinian children murdered out of sheer hatred, and who had destroyed his career and should never show his face in public again.

Others, however, called me something even worse than a genocidal maniac. They called me a Republican!

So I’d like to state for the record:

(1) In my opinion, Trump II remains by far the worst president in American history—beating out the second-worst, either Trump I or Andrew Jackson. Trump is destroying vaccines and science and universities and renewable energy and sane AI policy and international trade and cheap, lifesaving foreign aid and the rule of law and everything else that’s good, and he’s destroying them because they’re good—because even if destroying them hurts his own voters and America’s standing in the world, it might hurt the educated elites even more. It’s almost superfluous to mention that, while Trump himself is neither of these things, the MAGA movement that will anoint his successor now teems with antisemites and Holocaust “revisionists.”

(2) Thus, I’ll continue to vote straight-ticket Democrat, and donate money to Democrats, so long as the Democrats in question are seriously competing for Zionist Jewish votes at all—as, for example, has every Democratic presidential candidate in my lifetime so far.

(3) If it came down to an Israel-hating Squad Democrat versus a MAGA Republican, I’m not sure what I’d do, but I’d plausibly sit out the election or lodge a protest vote.

(4) In the extremely unlikely event that I had to choose between an Israel-hating Squad Democrat and some principled anti-MAGA Republican like Romney or Liz Cheney—then and only then do I expect that I’d vote Republican, for the first time in my life, a new and unfamiliar experience.

Scott Aaronson Staying sane on a zombie planet

Above is a typical sample of what’s been filling my inbox, all day every day. The emailers first ask me for reasoned dialogue—then, if I respond, they hit me with this stuff. I’m sharing because I think it’s a usefully accurate depiction of what several billion people, most academics in humanities fields, most who call themselves “on the right side of history,” and essentially all those attacking me genuinely believe about the world right now. Because of their anti-Nazism.

Hardly for the first time in my life, this weekend I got floridly denounced every five minutes—on SneerClub, on the blog of Peter Woit, and in my own inbox. The charge this time was that I’m a genocidal Zionist who wants to kill all Palestinian children purely because of his mental illness and raging persecution complex.

Yes, that’s right, I’m the genocidal one—me, whose lifelong dream is that, just like Germany and Japan rose from their necessary devastation in WWII to become pillars of our global civilization, so too the children in Gaza, the West Bank, Syria, Lebanon, and Iran will one day grow up in free and prosperous societies at peace with the West and with Israel. Meanwhile, those who demand an actual genocide of the Jews, another one—those who pray to Allah for it, who attempt it over and over, who preach it to schoolchildren, who celebrate their progress toward it in the streets—they’re all as innocent as lambs.

Yesterday, in The Free Press, came the report of a British writer who traveled to southern Lebanon, and met an otherwise ordinary young man there … who turned out to be excited for Muslims and Christians to join forces to slaughter all the Yahood, and who fully expected that writer would share his admiration for Hitler, the greatest Yahood-killer ever.

This is what the global far left has now allied itself with. This is what I’m right now being condemned for standing against, with commenter after commenter urging me to seek therapy.

To me, this raises a broader question: how exactly do you keep your sanity, when you live on a planet filled with brain-eaten zombies?

I’m still struggling with that question, but the best I’ve come up with is what I think of as the Weinberg Principle, after my much-missed friend and colleague here at UT Austin. Namely, I believe that it’s better to have one Steven Weinberg on your side while the rest of humanity is against you, than the opposite. Many other individuals (including much less famous ones) would also work here in place of Steve, but I’ll go with him because I think most of my readers would agree to three statements:

  1. Steve’s mind was more in sync with the way the universe really works, than nearly anyone else’s in history. He was to being free from illusions what Usain Bolt is to running or Magnus Carlsen is to chess.
  2. Steve’s toenail clippings constituted a greater contribution to particle physics than would the life’s work of a hundred billion Peter Woits.
  3. Steve’s commitment to Israel’s armed self-defense, and to Zionism more generally, made mine look weak and vacillating in comparison. No one need wonder what he would’ve said about Israel’s current war of survival against the Iranian-led terror axis.

Maybe it’s possible to wake the zombies up. Yoram Arnon, for example, wrote the following eloquent answer on Quora, in response to the question “Why are so many against freeing Palestine?”:

When Westerners think about freedom they think about freedom of speech, freedom of expression, freedom of movement, freedom of religion, freedom to form political parties, etc.

When Palestinians say “Free Palestine” they mean freedom from Jews, and from Israel’s existence. They’re advocating for the abolition of Israel, replacing it with an Arab country.

Israel is the only country in the Middle East that is free, in the Western sense of the word. If Israel were to disappear, Palestinians would fall under an autocratic regime, just like every other Arab country, with none of the above freedoms. And, of course, Israelis would suffer a terrible fate at their hands.

Pro Palestinians are either unable to see this, or want exactly that, but thankfully many in the West do see this – the same “many” that are against “freeing Palestine”.

Palestinians need to accept Israel’s right to exist, and choose to coexist peacefully alongside it, for them to have the peace and freedom the West wants for them.

Maybe reading words like these—or the words of Coleman Hughes, or Douglas Murray, or Hussein Aboubakr Mansour, or Yassine Meskhout, or John Aziz, or Haviv Rettig Gur, or Sam Harris, or the quantum computing pioneer David Deutsch—can boot a few of the zombies’ brains back up. But even then, I fear that these reboots will be isolated successes. For every one who comes back online, a thousand will still shamble along in lockstep, chanting “brainsssssss! genocide! intifada!”

I’m acutely aware of how sheer numbers can create the illusion of argumentative strength. I know many people who were sympathetic to Israel immediately after October 7, but then gradually read the room, saw which side their bread was buttered on, etc. etc. and became increasingly hostile. My reaction, of course, has been exactly the opposite. The bigger the zombie army I see marching against me, the less inclined I feel to become a zombie myself—and the clearer to me becomes the original case for the Zionist project.

So to the pro-Zionist students—Jewish of course, but also Christian, Muslim, Hindu, atheist, and everyone else—who feel isolated and scared to speak right up now, and who also often email me, here’s what I say. Yes, the zombies vastly outnumber us, but on the other hand, they’re zombies. Some of the zombies know longer words than others, but so far, not one has turned out to have a worldview terribly different from that of the image at the top of this post.


I’ll keep the comments closed, for much the same reasons I did in my last post.  Namely, while there are many people of all opinions and backgrounds with whom one can productively discuss these things, there are many more with whom one can’t. Furthermore, experience has shown that the latter can disguise themselves as the former for days on end, and thereby execute a denial-of-service attack on any worthwhile and open public discussion.

Addendum: The troll who sent the antisemitic image now says that he regrets and apologizes for it, and that he’s going to read books on Jewish history to understand his error. I’ll believe that when he actually sends me detailed book reports or other evidence, but just wanted to update.

September 02, 2025

Scott Aaronson Deep Gratitude

In my last post, I wrote about all the hate mail I’ve received these past few days. I even shared a Der-Stürmer-like image of a bloodthirsty, hook-nosed Orthodox Jew that some troll emailed me, after he’d repeatedly promised to send me a “diagram” that would improve my understanding of the Middle East. (Incredibly, commenters on Peter Woit’s blog then blamed me for this antisemitic image, mistakenly imagining that I’d created it myself, and then used their false assumption as further proof of my mental illness.)

Thanks to everyone who wrote to ask whether I’m holding up OK. The answer is: better than you’d expect! The first time you get attacked by dozens of Internet randos, it does feel like your life is over. But the sixth or seventh time? After you’ve experienced, firsthand, how illusory these people’s power over you actually is—how they can’t even dent your scientific career, can’t separate you from any of the friends who matter most to you (let alone your family), can’t really do anything to you beyond whatever they induce you to do to yourself? Then the deadly wolves appear more like poodles yapping from behind a fence. Try it and see!


Today I want to focus on a different kind of message that’s been filling my inbox. Namely, people telling me to stay strong, to keep up my courage, that everything I wrote strikes them as just commonsense morality.

It won’t surprise anyone that many of these people are Jews. But almost as many are not. I was touched to hear from several of my non-Jewish scientific colleagues—ones I’d had no idea were in my corner—that they are in my corner.

Then there was the American Gentile who emailed me a story about how, seeing an Orthodox family after October 7, he felt an urge to run up and tell them that, if worst ever came to worst, they could hide in his basement (“and I own guns,” he added). Amusingly, he added that his wife successfully dissuaded him from actually making such an offer, pointing out that it might freak out the recipients.

I replied that, here in America, I don’t expect that I’ll ever need to hide in anyone’s basement. But, I added, the only reason I don’t expect it is that there are so many Americans who, regardless of any religious or ideological differences, would hide their Jewish neighbors in their basements if necessary.

I also—despite neither I nor this guy exactly believing in God—decided to write a blessing for him, which came out as follows:

May your seed multiply a thousandfold, for like King Cyrus of Persia, you are a righteous man among the Gentiles.  But also, if you’re ever in Austin, be sure to hit me up for tacos and beer.


I’m even grateful, in a way, to SneerClub, and to Woit and his minions. I’m grateful to them for so dramatically confirming that I’m not delusional: some portion of the world really is out to get me. I probably overestimated their power, but not their malevolence.

I’ve learned, for example, that there are no words, however balanced or qualified, with which I can express the concept that Israel needs to defeat Hamas for the sake of both Israeli and Palestinian children, which won’t lead to Woit calling me a “genocide apologist who wants to see all the children in Gaza killed.” Nor are there any words with which to express my solidarity with the Jewish Columbia students who, according to an official university investigation, were last year systematically excluded from campus social life, intimidated, and even assaulted, and which won’t earn me names from Woit like “a fanatic allied with America’s fascist dictator.” Even my months-long silence about these topics got me labeled as “complicit with fascism and genocide.”

Realizing this is oddly liberating. When your back is to the wall in that way, either you can surrender, or else you can defend yourself. Your enemy has already done you the “favor” of eliminating any third options. Which, again, is just Zionism in a nutshell. It’s the lesson not only of 3,000 years of Jewish history, but also of superhero comics and of much of the world’s literature and cinema. It takes a huge amount of ideological indoctrination before such things stop being obvious.


Reading the SneerClubbers’ armchair diagnoses of my severe mental illness, paranoia, persecution complex, grandiosity, etc. etc. I had the following thought, paraphrasing Shaw:

Yes, they’re absolutely right that psychologically well-adjusted people generally do figure out how to adapt themselves to the reigning morality of their social environment—as indicated by the Asch conformity test, the Milgram electric-shock experiment, and the other classics of social psychology.

It takes someone psychologically troubled, in one way or another, to persist in trying to adapt the reigning morality of their social environment to themselves.

If so, however, this suggests that all the moral progress of humanity depends on psychologically troubled people—a realization for which I’m deeply grateful.

Jordan EllenbergThe Yankees’ #2 prospect

is a pitcher named Carlos Lagrange and if his major league nickname isn’t “The Multiplier” I will be sorely disappointed.

September 01, 2025

John BaezMariam Abu Dagga

Today I gave $10,000 to Doctors Without Borders, since they’re doing a lot of good work in Gaza. I made this gift in memory of Mariam Abu Dagga, a freelance photographer who was killed in the Nasser Hospital in the Gaza Strip on August 25th this year.

After the Israeli air force bombed this hospital, she and other journalists rushed in to check on the well-being of their colleague, Reuters journalist Hussam al-Masri. He had fact been killed.

In a ‘double tap’ attack, the Israeli air force then dropped more bombs, killing Dagga and others.

Wikipedia writes:

Dagga was known for documenting the experiences of displaced Palestinians, and of doctors who treated wounded or malnourished children, with “rare honesty and courage.” [2][3] Independent Arabia described Dagga as an “example of dedication and professional commitment,” bringing “her camera into the heart of the field.” [3] Associated Press Executive Director and Senior Vice President Julian Pace said that Dagga’s difficult journalistic work in Gaza was remarkable, particularly for her “coverage of the war’s impact on children.” [5]

Dagga won an internal award at the Associated Press for her coverage of malnourished children in Gaza. [5]

Speaking to the BBC, Palestinian journalist Hadar al-Qurd described Dagga as one of the most active female journalists in southern Gaza. [1] Al-Qurd said that other journalists relied on Dagga for the news and information that she gathered in her work. [1] Al-Qurd noted that Dagga was especially brave, always carrying her camera, and often going to the sites of airstrikes directly; she also noted that Dagga was a champion of the rights of female journalists in Gaza. [1] Other colleagues also noted Dagga’s bravery as a war correspondent. [3] Al Jazeera journalist Youmna El Sayed noted that Dagga had visibly lost a significant amount of weight during the war. [6]

Due to Israel’s repeated targeting and killing of journalists in the Gaza War, Dagga wrote her will during the conflict.

I urge you all to help Doctors Without Borders (= Médecines sans Frontiérs), who are doing good work in Gaza and other afflicted regions. They write:

MSF has over 1,300 staff working in Gaza’s hospitals, clinics, and other facilities, including our field hospital in Deir al-Balah and our clinic in Gaza City. Our teams provide surgical care, wound and burn care, malnutrition screening and treatment, maternal and pediatric care, physiotherapy, vaccination, mental health support, water and sanitation support, and care for non-communicable diseases, among other services. We are also providing rehabilitative care for war-wounded children we’ve managed to evacuate to our reconstructive surgery hospital in Amman, Jordan.

Facilities MSF has run or supported in Gaza

MSF teams have carried out lifesaving work in health facilities across Gaza, including in the Strip’s largest hospitals and our own clinics and field hospital. However, due to extremely volatile conditions on the ground—including attacks on health facilities and recurrent evacuation orders—our teams have had to move from facility to facility and continually adapt our activities. Hospitals and clinics we have supported during this war include:

• Nasser Hospital, European Gaza Hospital, and Martyrs, Beni Suhaila, Khan Younis, Al-Qarara, and Al-Attar clinics in Khan Younis
• Al-Aqsa Hospital and the MSF field hospital in Deir al-Balah    
• Al-Awda Hospital and mobile clinics in northern Gaza
• Al-Shifa Hospital, Al Helou Maternity Hospital, and our clinic in Gaza City   
• Rafah Indonesian Field Hospital, Emirati Maternity Hospital, Al-Najjar Hospital, and Al-Mawasi Health Post in Rafah

Water and sanitation activities

A lack of drinkable water, poor sanitation, and the destruction of water infrastructure have had dire consequences for people’s health in Gaza. Of more than 82,000 primary health care consultations MSF conducted in the first two months of this year, nearly a fifth were related to conditions linked with lack of water and hygiene, such as scabies and other skin conditions. This is why water distribution is an important part of MSF’s response.  

Between January and the end of April 2025, MSF teams distributed over 60 million liters of clean water and produced over 7.8 million liters through desalination. However, the already-dire water crisis in Gaza has worsened after Israeli authorities halted aid from entering the Strip on March 2, and then cut electricity on March 9, as water pumps and desalination plants require fuel and power to operate.

Humanitarian aid and medical supplies

Since Israeli authorities halted the flow of humanitarian aid and other supplies into the Strip on March 2, food, fuel, and medical stocks have been depleted. This total blockade of aid has deprived people of most basic needs and could lead to a high number of health complications and deaths.

Prior to the total siege, MSF provided over 636 tons of logistic and medical equipment from our international supply centers—as much as 30 planes or 130 trucks full. However, some supplies that are critical to our operations and the security of our staff have been difficult to transport into Gaza. These include generators, desalination stations and motor pumps, oxygen concentrators, vehicles, and equipment for communication.

Terence TaoA crowdsourced project to link up erdosproblems.com to the OEIS

Thomas Bloom’s erdosproblems.com site hosts nearly a thousand questions that originated, or were communicated by, Paul Erdős, as well as the current status of these questions (about a third of which are currently solved). The site is now a couple years old, and has been steadily adding features, the most recent of which has been a discussion forum for each individual question. For instance, a discussion I had with Stijn Cambie and Vjeko Kovac on one of these problems recently led to it being solved (and even formalized in Lean!).

A significantly older site is the On-line Encyclopedia of Integer Sequences (OEIS), which records hundreds of thousands of integer sequences that have some mathematician has encountered at some point. It is a highly useful resource, enabling researchers to discover relevant literature for a given problem so long as they can calculate enough of some integer sequence that is “canonically” attached to that problem that they can search for it in the OEIS.

A large fraction of problems in the Erdos problem webpage involve (either explicitly or implicitly) some sort of integer sequence – typically the largest or smallest size {f(n)} of some {n}-dependent structure (such as a graph of {n} vertices, or a subset of {\{1,\dots,n\}}) that obeys a certain property. In some cases, the sequence is already in the OEIS, and is noted in the Erdos problem web page. But in a large number of cases, the sequence either has not yet been entered into the OEIS, or it does appear but has not yet been noted on the Erdos web page.

Thomas Bloom and I are therefore proposing a crowdsourced project to systematically compute the hundreds of sequences associated to the Erdos problems and cross-check them against the OEIS. We have created a github repository to coordinate this process; as a by-product, this repository will also be tracking other relevant statistics about the Erdos problem website, such as the current status of formalizing the statements of these problems in the Formal Conjectures Repository.

The main feature of our repository is a large table recording the current status of each Erdos problem. For instance, Erdos problem #3 is currently listed as open, and additionally has the status of linkage with the OEIS listed as “possible”. This means that there are one or more sequences attached to this problem which *might* already be in the OEIS, or would be suitable for submission to the OEIS. Specifically, if one reads the commentary for that problem, one finds mention of the functions {r_k(N)} for {k=3,4,\dots}, defined as the size of the largest subset of {\{1,\dots,N\}} without a {k}-term progression. It is likely that several of the sequences {r_3(N)}, {r_4(N)}, etc. are in the OEIS, but it is a matter of locating them, either by searching for key words, or by calculating the first few values of these sequences and then looking for a match. (EDIT: a contributor has noted that the first foursequences appear as A003002, A003003, A003004, and A003005 in the OEIS, and the table has been updated accordingly.)

We have set things up so that new contributions (such as the addition of an OEIS number to the table) can be made by a Github pull request, specifically to modify this YAML file. Alternatively, one can create a Github issue for such changes, or simply leave a comment either on the appropriate Erdos problem forum page, or here on this blog.

Many of the sequences do not require advanced mathematical training to compute, and so we hope that this will be a good “citizen mathematics” project that can bring in the broader math-adjacent community to contribute to research-level mathematics problems, by providing experimental data, and potentially locating relevant references or connections that would otherwise be overlooked. This may also be a use case for AI assistance in mathematics through generating code to calculate the sequences in question, although of course one should always stay mindful of potential bugs or hallucinations in any AI-generated code, and find ways to independently verify the output. (But if the AI-generated sequence leads to a match with an existing sequence in the OEIS that is clearly relevant to the problem, then the task has been successfully accomplished, and no AI output needs to be directly incorporated into the database in such cases.)

This is an experimental project, and we may need to adjust the workflow as the project progresses, but we hope that it will be successful and lead to further progress on some fraction of these problems. The comment section of this blog can be used as a general discussion forum for the project, while the github issue page and the erdosproblems.com forum pages can be used for more specialized discussions of specific problems.

Tommaso DorigoSearching For Impossibly Rare Decays

I recently ran into a description of the Mu3e experiment, and got curious about it and the physics it studies. So after giving it a look, I am able to explain that shortly here - I think it is a great example of how deep our studies of particle physics are getting; or, on the negative side, how deep our frustration has gotten with the unassailable agreement of our experiments with Standard Model predictions.

Matter stable and unstable in the Standard Model

read more

August 30, 2025

Terence TaoSLMath announces new research programs

The Simons-Laufer Mathematical Sciences institute, or SLMath (formerly the Mathematical Sciences Research Institute, or MSRI) has recently restructured its program formats, and is now announcing three new research initiatives, whose applications open on Sep 1 2025:

  • AxIOM (Accelerating Innovation in Mathematics) is a new, month-long research program at SLMath, designed to accelerate innovation and introduce transformative ideas into the mathematical sciences. Programs begin in Spring 2027.
  • PROOF (Promoting Research Opportunities and Open Forums) is a two-week summer program designed to provide research opportunities for U.S.-based mathematicians, statisticians, and their collaborators in the U.S. and abroad, whose ongoing research may have been impacted by factors such as heavy teaching loads, professional isolation, limited access to funding, heavy administrative duties, personal obligations, or other constraints. Programs begin June-July 2026. The priority application deadline for PROOF 2026 is October 12, 2025.
  • Lasting Alliance Through Team Immersion and Collaborative Exploration (LATTICE) is a yearlong program which provides opportunities for U.S. mathematicians to conduct collaborative research on topics at the forefront of the mathematical and statistical sciences. Programs begin June-July 2026. LATTICE 2026 applications are open through February 1, 2026.

(Disclosure: I am vice-chair of the board of trustees at SLMath.)

n-Category Café Equivalence via Surjections

Pick a type of categorical structure: say bicategories, or monoidal categories, or whatever you like. Some of the functors between structures are equivalences, in whatever the appropriate sense might be. And some of those equivalences have one or both of these two properties:

  • They’re not just essentially surjective in every dimension — they’re actually surjective in every dimension.

  • They don’t just preserve the structure up to isomorphism or equivalence — they strictly preserve it.

Call an equivalence with both these properties a strict surjective equivalence. So a strict surjective equivalence is an equivalence of a very special and easy kind.

General principle: the standard notion of equivalence between structures is generated by just these very special ones. For example, two bicategories are biequivalent if and only if they can be linked up by a zigzag of strict surjective equivalences.

Why should we care? Because there are some types of structure where the right notion of equivalence isn’t clear, and this principle guides us to it. For example, it tells us the right notion of equivalence for double categories.

All this is done in my new paper:

Tom Leinster, Equivalence via surjections. arXiv:2508.20555, 2025.

I started thinking about this question during Maru Sarazola’s invited talk at Category Theory 2025 in Brno last month. She asked the question:

What is the right notion of equivalence between double categories?

and carefully went through the properties that the right notion of equivalence should have, some possible candidates, and different approaches one might take to deciding what “right” means.

The answer that Maru ultimately gave was that the right notion is “gregarious double equivalence”, proposed by Alexander Campbell in about 2020. And she gave a justification in terms of model categories, representing joint work between her, Lyne Moser and Paula Verdugo.

For the purposes of this post, it actually doesn’t matter what “gregarious double equivalence” means. What I want to talk about is the following principle, which popped into my head as Maru was speaking:

For many types of categorical structure, the natural notion of equivalence is generated, as an equivalence relation, by identifying AA and BB when there exists a strict surjective equivalence ABA \to B.

It occurred to me that this principle might give a rather different justification for why gregarious double equivalence is the right answer. And after some checking, I discovered that it does.

Let me explain.

A more concrete way to express the principle is that AA and BB are equivalent in the standard sense — whatever’s appropriate for the structures at hand — if and only if there exists a zigzag of strict surjective equivalences

A=A 0A 1A n=B. A = A_0 \leftarrow A_1 \rightarrow \ \cdots \ \leftarrow A_n = B.

For any type of categorical structure I can think of, the pullback of a strict surjective equivalence is a strict surjective equivalence. So a simpler concrete condition is just that there exists a span of strict surjective equivalences

ACB. A \leftarrow C \rightarrow B.

But hold on… what do I mean by “principle”?

What I mean is that for simple types of categorical structure, where “equivalence” and “strict surjective equivalence”, we have a theorem. Here are three examples.

  • Categories. We certainly know what it means for two categories to be equivalent. A “surjective equivalence” is an equivalence that’s not just essentially surjective on objects, but literally surjective on objects.

    In this case, the theorem is that categories AA and BB are equivalent if and only if there exists a span ACBA \leftarrow C \rightarrow B of surjective equivalences between them.

    (The word “strict” does nothing in this case.)

  • Monoidal categories. Again, we know what monoidal equivalence is, and it’s clear what a “strict surjective equivalence” is: a strict monoidal functor that’s a surjective equivalence of categories.

    The theorem is that monoidal categories AA and BB are monoidally equivalent if and only if there exists a span ACBA \leftarrow C \rightarrow B of strict surjective equivalences between them.

  • Bicategories. The pattern is the same. The standard notion of equivalence for bicategories is biequivalence. A “strict surjective equivalence”, in this setting, is a strict 22-functor that is literally surjective on objects and locally a surjective equivalence of categories. (Or put another way, surjective on 00-cells, locally surjective on 11-cells, and full and faithful on 22-cells.)

    The theorem is that bicategories AA and BB are biequivalent if and only if there exists a span ACBA \leftarrow C \rightarrow B of strict surjective equivalences between them.

Probably all these theorems are known. I included them in my paper because I couldn’t find them anywhere in the literature, not even the first one. But if you know a reference, I’d be glad to hear it.

Since the principle holds for categories, monoidal categories and bicategories, it’s reasonable to suppose that it might hold for other types of structure. And if we’re investigating some type of structure where the full notion of equivalence isn’t clear, this principle might help guide us to it.

For example, here’s a theorem on double categories, the main result of my paper:

  • Double categories. Again, it’s clear what “strict surjective equivalence” should mean: a strict double functor that’s surjective on 00-cells, locally surjective on both horizontal and vertical 11-cells, and full and faithful on 22-cells.

    The theorem is that double categories AA and BB are gregariously double equivalent if and only if there exists a span ACBA \leftarrow C \rightarrow B of strict surjective equivalences between them.

Even without me telling you what “gregarious double equivalence” means, the four theorems I’ve stated suggest that it’s the right notion of equivalence for double categories, because it continues the pattern we’ve seen for simpler categorical structures.

So, I agree with the conclusion that Moser, Sarazola and Verdugo had already reached! But for different reasons.

Incidentally, this must be the fastest paper I’ve ever written: just under six weeks from sitting in Maru’s talk and hearing the mathematical term “gregarious” for the first time ever to putting the paper on the arXiv. But the principle that all equivalences are generated by strict surjective equivalences was planted in my head in the late 1990s or early 2000s by Carlos Simpson. Back then, we were both working on higher category theory, and when he explained this principle, I found it very striking — so striking that I remembered it 20+ years later. There’s a bit more on that higher categorical context in the introduction to my paper.

August 29, 2025

Matt von HippelTwo Types of Scientific Fraud: for a Fee and for Power

A paper about scientific fraud has been making the rounds in social media lately. The authors gather evidence of large-scale networks of fraudsters across multiple fields, from teams of editors that fast-track fraudulent research to businesses that take over journals, sell spots for articles, and then move on to a new target when the journal is de-indexed. I’m not an expert in this kind of statistical sleuthing, but the work looks impressively thorough.

Still, I think the authors overplay their results a bit. They describe themselves as revealing something many scientists underestimate. They point to what they label as misconceptions: that scientific fraud is usually perpetrated alone by individual unethical scientists, or that it is almost entirely a problem of the developing world, and present their work as disproving those misconceptions. Listen to them, and you might get the feeling that science is rife with corruption, that no result, or scientist, can be trusted.

As far as I can tell, though, those “misconceptions” they identify are true. Someone who believes that scientific fraud is perpetrated by loners is probably right, as is someone who believes it largely takes place outside of the first world.

As is often the case, the problem is words.

“Scientific Fraud” is a single term for two different things. The two both involve bad actors twisting scientific activity. But in everything else — their incentives, their geography, their scale, and their consequences — they are dramatically different.

One of the types of scientific fraud is largely about power.

In references 84-89 of the paper, the authors give examples of large-scale scientific fraud in Europe and the US. All (except one, which I’ll mention later) are about the career of a single researcher. Each of these people systematically bent the truth, whether with dodgy statistics, doctored images, or inflating citation counts. Some seemed motivated to promote a particular scientific argument, cutting corners to push a particular conclusion through. Others were purer cases of self-promotion. These people often put pressure on students, postdocs, and other junior researchers in their orbits, which increases the scale of their impact. In some cases, their work rippled out to convince other researchers, prolonging bad ideas and strangling good ones. These were people with power, who leveraged that power to increase their power.

There also don’t appear to be that many of them. These people are loners in a meaningful sense, cores of fraud working on their own behalf. They don’t form networks with each other, for the most part: because they work towards their own aggrandizement, they have no reason to trust anyone else doing the same. I have yet to see evidence that the number of these people is increasing. They exist, they’re a problem, they’re important to watch out for. But they’re not a crisis, and they shouldn’t shift your default expectations of science.

The other, quite different, type of scientific fraud is fraud for a fee.

The cases this paper investigates seem to fall into this category. They are businesses, offering the raw material of academic credit (papers, co-authorship, citations, publication) for cash. They’re paper mills, of various sorts. These are, at least from an academic perspective, large organizations, with hundreds or thousands of customers and tens of suborned editors or scientists farming out their credibility. As the authors of this paper argue, fraudsters of this type are churning out more and more papers, potentially now fueled by AI, adding up to a still small, but non-negligible, proportion of scientific papers in total.

Compared to the first type of fraud, though, buying credit in this way doesn’t give very much power. As the paper describes, many of the papers churned out by paper mills don’t even go into relevant journals: for example, they mention “an article about roasting hazelnuts in a journal about HIV/AIDS care”. An article like that isn’t going to mislead the hazelnut roasting community, or the HIV/AIDS community. Indeed, that would be counter to its purpose. The paper isn’t intended to be read at all, and ideally gets ignored: it’s just supposed to inflate a number.

These numbers are most relevant in the developing world, and when push comes to shove, almost all of the buyers of these services identified by the authors of this paper come from there. In many developing countries, a combination of low trust and advice from economists leads to explicit point systems, where academics are paid or hired explicitly based on criteria like where and how often they publish or how they are cited. The more a country can trust people to vouch for each other without corruption, the less these kinds of incentives have purchase. Outside of the developing world, involvement in paper mills and the like generally seems to involve a much smaller number of people, and typically as sellers, not buyers: selling first-world credibility in exchange for fees from many developing-world applicants.

(The one reference I mentioned above is an interesting example of this: a system built out of points and low trust to recruit doctors from the developing world to the US, gamed by a small number of co-authorship brokers.)

This kind of fraud doesn’t influence science directly. Its perpetrators aren’t trying to get noticed, but to keep up a cushy scam. You don’t hear their conclusions in the press, other scientists don’t see their work. Instead, they siphon off resources: cannibalizing journals, flooding editors with mass-produced crap, and filling positions and slurping up science budgets in the countries that can least afford them. As they publish more and more, they shouldn’t affect your expectations of the credibility of science: any science you hear about will be either genuine, or fraud from the other category. But they do make the science you hear about harder and harder to do.

(The authors point out one exception: what about AI? If a company trains a large language model on the current internet, will its context windows be long enough to tell that that supposedly legitimate paper about hazelnuts is in an HIV/AIDS journal? If something gets said often enough, copied again and again in papers sold by a mill, will an AI trained on all these papers be convinced? Presumably, someone is being paid good money to figure out how to filter AI-generated slop from training data: can they filter paper mill fraud as well?)

It’s a shame that we have one term, scientific fraud, to deal with these two very different things. But it’s important to keep in mind that they are different. Fraud for power and fraud for money can have very different profiles, and offer very different risks. If you don’t trust a scientific result, it’s worth understanding what might be at play.

Scott Aaronson Deep Zionism

Suppose a man has already murdered most of your family, including several of your children, for no other reason than that he believes your kind doesn’t deserve to exist on earth. The murderer was never seriously punished for this, because most of your hometown actually shared his feelings about your family. They watched the murders with attitudes ranging from ineffectual squeamishness to indifference to unconcealed glee.

Now the man has kidnapped your last surviving child, a 9-year-old girl, and has tied her screaming to train tracks. You can pull a lever to divert the train and save your daughter. But there’s a catch, as there always is in these moral dilemmas: namely, the murderer has also tied his own five innocent children to the tracks, in such a way that, if you divert the train, then it will kill his children. What’s more, the murderer has invited the entire town to watch you, pointing and screaming “SHAME!!” as you agonize over your decision. He’s persuaded the town that, if you pull the lever, then having killed five of his children to save only one of yours, you’re a far worse murderer than he ever was. You’re so evil, in fact, that he’s effectively cleansed of all guilt for having murdered most of your family first, and the town is cleansed of all guilt for having cheered that. Nothing you say can possibly convince the town otherwise.

The question is, what do you do?

Zionism, to define it in one sentence, is the proposition that, in the situation described, you have not merely a right but a moral obligation to pull the lever—and that you can do so with your middle finger raised high to the hateful mob. Zionism is the belief that, while you had nothing against the murderer’s children, while you would’ve wanted them to grow up in peace and happiness, and while their anguished screams will weigh on your conscience forever, as your children’s screams never weighed on the murderer’s conscience, or on the crowd’s—even so, the responsibility for those children’s deaths rests with their father for engineering this whole diabolical situation, not with you. Zionism is the idea that the correct question here is the broader one: “which choice will bring more righteousness into the world, which choice will better embody the principle that no one’s children are to be murdered going forward?” rather than the narrowly utilitarian question, “which choice will lead to fewer children getting killed right this minute?” Zionism is the conviction that, if most of the world fervently believes otherwise, than most of the world is mistaken—as the world has been mistaken again and again about the biggest ethical questions all through the millennia.

Zionism, so defined, is the deepest moral belief that I have. It’s deeper than any of my beliefs about “politics” in the ordinary sense. Ironically, it’s even deeper than my day-to-day beliefs about the actual State of Israel and its neighbors. I might, for example, despise Benjamin Netanyahu and his ministers, might consider them incompetent and venal, might sympathize with the protesters who’ve filled the streets of Tel Aviv to demand their removal. Even so, when the murderer ties my child to the train tracks and the world cheers the murderer on, not only will I pull the lever myself, I’ll want Benjamin Netanyahu to pull the lever if he gets to it first.

Crucially, everything worthwhile in my life came when, and only when, I chose to be “Zionist” in this abstract sense: that is, steadfast in my convictions even in the face of a jeering mob. As an example, I was able to enter college three years early, which set the stage for all the math and science I later did, only because I finally said “enough” to an incompetent school system where I was bullied and prevented from learning, and to teachers and administrators whose sympathies lay with the bullies. I’ve had my successes in quantum computing theory only because I persisted in what at the time was a fairly bizarre obsession, rather than working on topics that almost everyone around me considered safer, more remunerative, and more sensible.

And as the world learned a decade ago, I was able to date, get married, and have a family, only because I finally rejected what I took to be the socially obligatory attitude for male STEM nerds like me—namely, that my heterosexuality was inherently gross, creepy, and problematic, and that I had a moral obligation never to express romantic interest to women. Yes, I overestimated the number of people who ever believed that, but the fact that it was clearly a nonzero number had been deterrent enough for me. Crucially, I never achieved what I saw for years as my only hope in life, to seek out those who believed my heterosexuality was evil and argue them out of their belief. Instead I simply … well, I raised a middle finger to the Andrea Dworkins and Arthur Chus and Amanda Marcottes of the world. I went Deep Zionist on them. I asked women out, and some of those women (not having gotten the memo that I was “problematic,” gross, and worthless) said yes, and one of them became my wife and the mother of my children.

Today, because of the post-October-7 public stands I’ve taken in favor of Israel’s continued existence, I deal with emails and social media posts day after day calling me a genocidal baby-killing monster. I’ve lost perhaps a dozen friends (while retaining hundreds more friends, and gaining some new ones). The haters’ thought appears to be that, if they can just raise the social cost high enough, I’ll finally renounce my Zionist commitments and they can notch another win. In this, they oddly mirror Hamas, Hezbollah, and the IRGC, who think that, if they can just kill and maim enough Israelis, the hated “settler-colonialist rats” will all scurry back to Poland or wherever else they came from (best not to think too hard about where they did come from, what was done to them in those places, how the Palestinian Arabs of the time felt about what was done to them, or how the survivors ended up making a last stand in their ancestral home of Israel—even afterward, repeatedly holding out olive branches that were met time after time with grenades).

Infamously, Israel’s enemies have failed to understand for a century that, the more they rape and murder, the more Zionist the hated Zionists will become, because unlike the French in Algeria or whatever, most of the Zionists have no other land to go back to: this is it for them. In the same way, my own haters don’t understand that, the more they despise me for being myself, the more myself I’ll be, because I have no other self to turn into.

I’m not opening the comments on this post, because there’s nothing here to debate. I’m simply telling the world my moral axioms. If I wrote these words, then turned to pleading with commenters who hated me because of them, then I wouldn’t really have meant the words, would I?

To my hundreds of dear friends and colleagues who’ve stood by me the past two years, to the Zionists and even just sympathetic neutrals who’ve sent me countless messages of support, but who are too afraid (and usually, too junior in their careers) to speak up in public themselves: know that I’ll use the protections afforded by my privileged position in life to continue speaking on your behalf. Know that I’m infinitely grateful, that you give me strength, and that if I can give you a nanoparticle of strength back to you, then my entire life wasn’t in vain. And if I go silent on this stuff from time to time, for the sake of my mental health, or to spend time on quantum computing research or my kids or the other things that bring me joy—never take that to mean that I’ve capitulated to the haters.

To the obsessive libelers, the Peter Woits and other snarling nobodies, the self-hating Jews, and those who’d cheer to see Israel “decolonized” and my friends and family there murdered, I say—well, I don’t say anything; that’s the point! This is no longer a debate; it’s a war, and I’ll simply stand my ground as long as I’m able. Someday I might forgive the Gentiles among you if you ever see the light, if you ever realize how your unreflective, social-media-driven “anti-fascism” led you to endorse a program that leads to the same end as the original Nazi one. The Jews among you I’ll never forgive, because you did know better, and still chose your own comfort over the physical survival of your people.

It might as well be my own hand on the madman’s lever—and yet, while I grieve for all innocents, my soul is at peace, insofar as it’s ever been at peace about anything.


Update (Aug. 29): This post was born of two years of frustration. It was born of trying, fifty or a hundred times since October 7, to find common ground with the anti-Zionists who emailed me, messaged me, etc.—“hey, obviously neither of us wants any children killed or starved, we both have many bones to pick with the current Israeli government, but surely we at least agree on the necessity of defeating Hamas, right? right??“—only to discover, again and again, that the anti-Zionists had no interest in such common ground. With the runaway success of the global PR campaign against Israel—i.e., of Sinwar’s strategy—and with the rise of figures like Mamdani (and his right-wing counterparts) all over the Western world, anti-Zionists smell blood in the water today. And so, no matter how reasonable they presented themselves at first, eventually they’d come out with “why can’t the Jews just go back to Germany and Poland?” or “the Holocaust was just one more genocide among many; it doesn’t deserve any special response,” or “why can’t we dismantle Israel and have a secular state, with a Jewish minority and a majority that’s sworn to kill all Jews as soon as possible?” And then I realize, with a gasp, that we Jews really are mostly on our own in a cruel and terrifying world—just like we’ve been throughout history.

To say that this experience radicalized me would be an understatement. Indeed, my experience has been that even most Israelis, who generally have far fewer illusions than we diaspora Jews, don’t understand the vastness of the chasm that’s formed. They imagine that they can have a debate with outsiders similar to the debates playing out within Israel—one that presupposes basic factual knowledge and the parameters of the problem (e.g., clearly we can’t put 7 million Jews under the mercy of Hamas). The rationale for Zionism itself feels so obvious to them as to be cringe. Except that, to the rest of the world, it isn’t.

We’re not completely on our own though. There remain decent people of every background, who understand the stakes and feel the weight of history—and I regularly hear from them. And whatever your criticisms of Israel’s current tactics, so long as you accept the almost comically overwhelming historical case for the necessity of Jewish self-defense, this post wasn’t aimed at you, and you and I probably could discuss these matters. It’s just that the anti-Zionists scream so loudly, suck up so much oxygen, that we definitely can’t discuss them in public. Maybe in person sometime, face to face.

John BaezRupert’s Property



You can cut a hole in a cube that’s big enough to slide an identical cube through that hole! Think about that for a minute—it’s kind of weird.

Amazingly, nobody could prove any convex polyhedron doesn’t have this property! It’s called ‘Rupert’s property’.

Until this week.

This week Steininger and Yurkevich proved there is a convex polyhedron that you can’t cut a hole in big enough to slide the entire polyhedron through the hole. It has 90 vertices, and apparently 240 edges and 152 faces.



To prove that no such hole is possible, they had to do a computer search of 18 million different holes, plus use a lot of extra math to make sure they’d checked enough possibilities:

• Jakob Steininger and Sergey Yurkevich, A convex polyhedron without Rupert’s property.

To celebrate their discovery, they gave this polyhedron a silly name. Since this polyhedron lacks Rupert’s property, they called it a ‘noperthedron’.

Why is this property called ‘Rupert’s property’? Wikipedia explains:

In geometry, Prince Rupert’s cube is the largest cube that can pass through a hole cut through a unit cube without splitting it into separate pieces. Its side length is approximately 1.06, 6% larger than the side length 1 of the unit cube through which it passes. The problem of finding the largest square that lies entirely within a unit cube is closely related, and has the same solution.

Prince Rupert’s cube is named after Prince Rupert of the Rhine, who asked whether a cube could be passed through a hole made in another cube of the same size without splitting the cube into two pieces. A positive answer was given by John Wallis. Approximately 100 years later, Pieter Nieuwland found the largest possible cube that can pass through a hole in a unit cube.

Here Greg Egan shows how Rupert’s property works for the cube:

Here he shows how it works for the regular octahedron:

And finally, here’s a video by David Renshaw showing 26 polyhedra with Rupert’s property… and 5 polyhedra that might lack it:

 

The triakis tetrahedron is an extremely close call, but it does have Rupert’s property:

August 28, 2025

Doug Natelson25 years of Nano Letters

Back in the dawn of the 21st century, the
American Chemical Society founded a new journal, Nano Letters, to feature letters-length papers about nanoscience and nanotechnology.  This was coincident with the launch of the National Nanotechnology Initiative, and it was back before several other publishers put out their own nano-focused journals.  For a couple of years now I've been an associate editor at NL, and it was a lot of fun to work with my fellow editors on putting together this roadmap, intended to give a snapshot of what we think the next quarter century might hold.  I think some of my readers will get a kick out of it.  

August 27, 2025

John BaezPolycyclic Aromatic Hydrocarbons

In 2004, a team of scientists discovered hydrocarbons called anthracene and pyrene in an amazing structure called the Red Rectangle!

Here two stars 2300 light years from us are spinning around each other while pumping out a huge torus of icy dust grains and hydrocarbon molecules. It’s not really shaped like a rectangle or X—it just looks that way from here. The whole scene is about one third of a light year across.

This was first time such complex molecules had been found in space:

• Uma P. Vijh, Adolf N. Witt, and Karl D. Gordon, Small polycyclic aromatic hydrocarbons in the Red Rectangle, The Astrophysical Journal 619 (2005), 368–378.

Wherever carbon-containing materials suffer incomplete combustion, you get PAHs, or polycyclic aromatic hydrocarbons. In Earth you can find them in soot, or the tarry stuff that forms in a barbecue grill. They’re common in outer space, too.

But what are PAHs like, exactly? They’re made of hexagonal rings of carbon atoms, with some hydrogens along the edges:

Benzene has a single hexagonal ring, with 6 carbons and 6 hydrogens. You’ve probably heard about naphthalene, which is used for mothballs: this consists of two hexagonal rings stuck together. True PAHs have more. With three rings you can make anthracene:

and phenanthrene:

With four, you can make napthacene:

pyrene:

triphenylene:

and chrysene:

And so on! The game just gets more complicated as you get to use more puzzle pieces.

By now, lots of organic molecules have been found in interstellar or circumstellar space. There’s a whole ‘ecology’ of organic chemicals out there, engaged in complex reactions. Life on planets might someday be seen as just an aspect of this larger ecology. PAHs are probably among the dominant players in this ecology, at least at this stage.

Indeed, I’ve read that about 10% of the interstellar carbon is in the form of PAHs—big ones, with about 50 carbons per molecule. They’re common because they’re incredibly stable. They’ve even been found riding the shock wave of a supernova explosion!

PAHs are also found in meteorites called ‘carbonaceous chondrites’. These space rocks contain just a little carbon: about 3% by weight. But 80% of this carbon is in the form of PAHs.

Here’s an interview with a scientist who thinks PAHs were important precursors of life on Earth:

Aromatic world, interview with Pascale Ehrenfreund, Space Daily.

Also try this:

PAH world hypothesis, Wikipedia.

August 26, 2025

Tommaso DorigoA Remarkable Graph: The Full Dalitz Plot Of Neutron Decay

The neutron is a fascinating particle, and one which has kept experimental physicists busy for almost a century now. Discovered by James Chadwick in 1932 in a cunning experiment which deserves a separate post (it is a promise, or a threat if you prefer),  the neutron has been all along a protagonist in the development of nuclear weapons as well as in the extraction of nuclear power from fission reactors. And of more relevance to our discussion here, it has powered endless studies both in the context of nuclear and subnuclear physics.

read more

August 24, 2025

August 23, 2025

Peter Rohde Photo albums

Peter’s photos: https://www.icloud.com/sharedalbum/#B275oqs3qKSZvQ

Screenshots: https://www.icloud.com/sharedalbum/#B27532ODWjIQb9

Climbing book launch: https://www.icloud.com/sharedalbum/#B27GWZuqDGnuOyN

Salisbury waters: https://www.icloud.com/sharedalbum/#B275qXGF1JQFkx

Christmas with Ash: https://www.icloud.com/sharedalbum/#B27G6XBubAhoT6

Hosin BBQ duck: https://www.icloud.com/sharedalbum/#B27GY8gBYG3b5mD

Hawks Nest to Smiths Lake: https://www.icloud.com/sharedalbum/#B2759UlCqSH5bE

Europe & Alps: https://www.icloud.com/sharedalbum/#B275ON9t3W0lu

Point Perpendicular: https://www.icloud.com/sharedalbum/#B27GqkRUiGivXD2

Newnes canyoning: https://www.icloud.com/sharedalbum/#B27GfnH8tgHSmX

Coffs Harbour to Yamba: https://www.icloud.com/sharedalbum/#B27J0DiRHJKuuWr

Wendy Bruere Christmas (2020): https://www.icloud.com/sharedalbum/#B27G4TcsmGoHysj

Six Foot Track: https://www.icloud.com/sharedalbum/#B2753qWtHZA9EX

Kosciusko to Kiandra: https://www.icloud.com/sharedalbum/#B27GgZLKuGaewVm

Camping food: https://www.icloud.com/sharedalbum/#B27GtnIORgbmHu

The Aardvark: https://www.icloud.com/sharedalbum/#B275VaUrzvmAiT

Kangaroo Valley kayaking: https://www.icloud.com/sharedalbum/#B27JEsNWnJrCpi0

Claustral canyon: https://www.icloud.com/sharedalbum/#B2755Z2WMOTpsk

Budawang: https://www.icloud.com/sharedalbum/#B27GDdyTvGvpINL

Mother’s Day panoramas (2021): https://www.icloud.com/sharedalbum/#B27GFssfGG9WmJP

Point Perpendicular & Nowra: https://www.icloud.com/sharedalbum/#B27GRMtznGPdeuZ

Blood moon: https://www.icloud.com/sharedalbum/#B27GdIshaG8NgGX

La Perouse to Coogee: https://www.icloud.com/sharedalbum/#B275aVbMK4h7qo

Canberra ASPI launch: https://www.icloud.com/sharedalbum/#B27GQOeMmGj4Zcv

Edible foraging: https://www.icloud.com/sharedalbum/#B275ejO179Si0N

Sydney to Wollongong: https://www.icloud.com/sharedalbum/#B275M7GFPUasMe

Album for Dad, Father’s Day (2021): https://www.icloud.com/sharedalbum/#B2752plgjnnkUe

Vaucluse (with Cheryl, Nestor & Wendy): https://www.icloud.com/sharedalbum/#B275CmvAS4uA0Z

Bouddi National Park: https://www.icloud.com/sharedalbum/#B27GdPblXG8WdOo

Tom Thumb (the 2nd): https://www.icloud.com/sharedalbum/#B275aDWbr4CN2w

Eden to Victoria: https://www.icloud.com/sharedalbum/#B27GJDfWGArX8l

Wendy’s book launch (the 2nd): https://www.icloud.com/sharedalbum/#B27GIcgc2G7h08y

Mark & Pat Bruere visit Sydney: https://www.icloud.com/sharedalbum/#B27G0ehgLbyWyg

New Years Eve climb (2021): https://www.icloud.com/sharedalbum/#B27Ju8EH6JOZxmU

Newnes Canyoning (2022): https://www.icloud.com/sharedalbum/#B275BydzFU0GZ8

Royal National Park (2022): https://www.icloud.com/sharedalbum/#B27GlxzuqGVI5nE

Peter & Wendy: https://www.icloud.com/sharedalbum/#B27Gf693ZG52tfd

Book photo shoots: too rude…

Wendy & Peter’s mushroom trip: https://www.icloud.com/sharedalbum/#B27GrhkPxG27So8

Post-mushroom hike: https://www.icloud.com/sharedalbum/#B27GdFryYG8i3Ur

Wendy Kalymnos favourites: https://www.icloud.com/sharedalbum/#B27JqstnBJEXkH2

Wendy Frenchmans screenshots: https://www.icloud.com/sharedalbum/#B27Jr1PPdJpd7Dq

Instagram: https://www.icloud.com/sharedalbum/#B27GzFCC1Gb4tqr

Haute route: https://www.icloud.com/sharedalbum/#B27J8GySPJtWoQ1

Kim’s KKKalendar: https://www.icloud.com/sharedalbum/#B275fk75vIL0sH

Frenchmans Cap Wild: https://www.icloud.com/sharedalbum/#B27G4VTwGGoFBkz

Photoshoot with Zixin: https://www.icloud.com/sharedalbum/#B27GPCdxkGKPkM4

Wendy birthday hike (2023): https://www.icloud.com/sharedalbum/#B27GWBC59GnHpQW

Bateman’s Bay to Bawley Point: https://www.icloud.com/sharedalbum/#B27JsHvHoJ8bxWf

Stockton Sand dunes (2023): https://www.icloud.com/sharedalbum/#B27GVfZ2vGloFZV

Wendy book launch (2023): https://www.icloud.com/sharedalbum/#B27J058xyJR4IBM

Dolomites (2023): https://www.icloud.com/sharedalbum/#B0Z5kuVsbGJUzKO

Mount Arapiles: https://www.icloud.com/sharedalbum/#B275GH8Mq8Uh2X

Mount Solitary loop: https://www.icloud.com/sharedalbum/#B275nhQST2mETE

Klaus Hanz Franz Rohde Kunst: https://www.icloud.com/sharedalbum/#B27GqQrCLGiY3vb

Klaus Rohde funeral slideshow: https://www.icloud.com/sharedalbum/#B27GDZLe8GXP58K

Dad (old, B&W): https://www.icloud.com/sharedalbum/#B27GLLXGLJ5mbT2

Klaus & Ursula wedding: https://www.icloud.com/sharedalbum/#B275cLqfN7154g

Test Greece: https://www.icloud.com/sharedalbum/#B27Jq4WnLJ6JMNd

From Will Skea (Alps): https://www.icloud.com/sharedalbum/#B27JHciePJFwacG

From Will Skea (Frenchmans Cap): https://www.icloud.com/sharedalbum/#B275ZhN2v3EVq6

From Will Skea (Arapiles): https://www.icloud.com/sharedalbum/#B27JPrgBGJu3BTD

Coffs Harbour to Yamba (2): https://www.icloud.com/sharedalbum/#B27GFqhgJG9LHgT

Mark magic show (2021): https://www.icloud.com/sharedalbum/#B27G60dj6ARCvd

Wendy Christmas present (2020): https://www.icloud.com/sharedalbum/#B275FrPQ6GxvRu

AHS 25 year reunion: https://www.icloud.com/sharedalbum/#B275O3DjHUvSv

WhatsApp: https://www.icloud.com/sharedalbum/#B275tzEA5fX1nc

Armidale High School: https://www.icloud.com/sharedalbum/#B27GnbeumG4PnAF

Book photos for Mum & Dad: https://www.icloud.com/sharedalbum/#B27Gtec4XQkASe

Miscellaneous: https://www.icloud.com/sharedalbum/#B27Gq6kMgGKn7GR

Three Capes Trail (2022): https://www.icloud.com/sharedalbum/#B27G7HOIlGrDUGZ

Childhood computer programming: https://www.icloud.com/sharedalbum/#B275fu2MutDU8N

Magic with Mark in Maroubra: https://www.icloud.com/sharedalbum/#B27Gv6DhEGD9U3G

Photoshoot with Zixin (2024): https://www.icloud.com/sharedalbum/#B27GCATCnJGoRfW

Butt Crack (2021): https://www.icloud.com/sharedalbum/#B275VtHQfMv0zw

Greece photos new (edited to remove photos from wrong album): https://www.icloud.com/sharedalbum/#B27GY3uThGoBcGj

Singapore (all combined): https://www.icloud.com/sharedalbum/#B275qsTcwJKJjl

Hong Kong (transit): https://www.icloud.com/sharedalbum/#B2759v1AbS8Hve

Taiwan: https://www.icloud.com/sharedalbum/#B27GQD2D7Gw0hAp

India (combined): https://www.icloud.com/sharedalbum/#B27Gtue8VQy83g

Freycinet: https://www.icloud.com/sharedalbum/#B27G5VpecGE5Tbg

Triglav: https://www.icloud.com/sharedalbum/#B275MbK9Vy8erz

Shared with me: https://www.icloud.com/sharedalbum/#B27GGXqixzPOrm

Mount Wellington climbing: https://www.icloud.com/sharedalbum/#B27Gd59qiG8Kjy4

New Zealand combined (2004): https://www.icloud.com/sharedalbum/#B27GIZ8BIGNN5jy

New Zealand combined (2005): https://www.icloud.com/sharedalbum/#B27GcuRfIGFVIcL

Yea: https://www.icloud.com/sharedalbum/#B27GZYbYHGhFIir

Mount Pleasant: https://www.icloud.com/sharedalbum/#B275Iy2hC0JTTL

D’Aguilar: https://www.icloud.com/sharedalbum/#B27Gh7fzTGZBosS

Bali (2001): https://www.icloud.com/sharedalbum/#B27G1qNHBGOTbIr

Samba Ninjas: https://www.icloud.com/sharedalbum/#B27GG34bAzqQ0v

Armidale (misc): https://www.icloud.com/sharedalbum/#B27GSkLVwGyobbX

Emma’s party (2008): https://www.icloud.com/sharedalbum/#B275S2ms99Zyby

Goettingen (2011): https://www.icloud.com/sharedalbum/#B27JIrbT3Jsgxhd

South Coast track: https://www.icloud.com/sharedalbum/#B27G58NWBG6QyN7

Childhood (misc): https://www.icloud.com/sharedalbum/#B27GVU1CZGmfOcg

Minsk (2006): https://www.icloud.com/sharedalbum/#B27G3JpSBGX1UkQ

Baden-Baden (2019): https://www.icloud.com/sharedalbum/#B27595X5HTVzJr

Berlin (combined): https://www.icloud.com/sharedalbum/#B27JqWzChJ6qizD

Switzerland (combined): https://www.icloud.com/sharedalbum/#B275zXwoYGJ6HMF

Italy highlights: https://www.icloud.com/sharedalbum/#B27G47PHQGoJium

Germany (misc): https://www.icloud.com/sharedalbum/#B275hPMfYGu5xVJ

Garmisch (2022): https://www.icloud.com/sharedalbum/#B27GFsbvlG9Xrr6

Germany (2019): https://www.icloud.com/sharedalbum/#B27G6Mn98G56Ncb

Garmisch (2006): https://www.icloud.com/sharedalbum/#B27J5lIdKGLC9KG

Baden-Baden (2005): https://www.icloud.com/sharedalbum/#B275sWRpHHQkt9

Berlin (2005): https://www.icloud.com/sharedalbum/#B27GgOQtrGjQrpH

Zugspitze (2005): https://www.icloud.com/sharedalbum/#B27G81mNdGcApGt

Amsterdam, Bristol (2006): https://www.icloud.com/sharedalbum/#B275B9SRzyBjlH

Baden-Baden (2006): https://www.icloud.com/sharedalbum/#B275eD9V79I2XR

Berlin (2006): https://www.icloud.com/sharedalbum/#B275toRf1fH8MD

Berlin, Jena (2007): https://www.icloud.com/sharedalbum/#B27GTI3fvGVgNit

Erlangen (2006): https://www.icloud.com/sharedalbum/#B27JrotZ2JpMb0i

Garmisch (2010): https://www.icloud.com/sharedalbum/#B27JPJPSiJurzNg

Germany (2010): https://www.icloud.com/sharedalbum/#B275FhYPQP650

Stuttgart (2006): https://www.icloud.com/sharedalbum/#B27GmitydGVVaZh

Changi (2019): https://www.icloud.com/sharedalbum/#B27GnmlKoG4JHpX

Japan (2007): https://www.icloud.com/sharedalbum/#B275AerZbG6FxVL

Japan (2012): https://www.icloud.com/sharedalbum/#B27GjBjobGg6PUa

Miscellaneous (including Japan 2013): https://www.icloud.com/sharedalbum/#B27GTpbybGySbE8

Currumbin & Tugin (2021): https://www.icloud.com/sharedalbum/#B275vBKZ4xH9X6

Brisbane (2021): https://www.icloud.com/sharedalbum/#B275YHsSjxQnm0

Weed in Byron (26/6/2025): https://www.icloud.com/sharedalbum/#B275Q2ydoGsQ4O5

Weed in Byron 2: https://www.icloud.com/sharedalbum/#B27GQDYhLGwsuY4

August 22, 2025

Peter Rohde Why?

  1. The person dressed up as Ursula pretending to be my mother clearly isn’t and hasn’t been for a long time.
  2. When I went back to Armidale after leaving BTQ and being left unemployed she made numerous ongoing promises to provide me with assistance, both in obtaining my own accommodation and providing financial assistance.
  3. These didn’t materialise and the promises were revoked.
  4. Instead I was evicted from the family home and subject to ongoing stalking and harassment that required multiple referrals to law enforcement, both to the police and the Attorney-General, demanding cease and desist.
  5. These have been systematically ignored and up until the last message she continues to bypass these requests, approaching my personal friends to harass me and stalk me indirectly. The messages passed on are the usual fake “I’m worried about him” bullshit.
  6. Why has my family home been confiscated by security, who actively break the law by ignoring cease and desist from stalking notices made to law enforcement, forcing an unemployed civilian into ongoing homelessness since early in the year?
  7. What is the rational for my eviction and being barricaded from my own home?
  8. I continue to face a medical blockade and am unable to access essential medicines. Seroquel scripts are deliberately delayed past known script deadlines to try and destabilise me.
  9. Vyvanse scripts are denied outright as the psychiatrist does not respond. He is also known to be a state actor.
  10. It has been repeatedly indicated to me not to worry about finances because they have my back. Instead now the only cash I have is that obtained from fully drawing out a cash advance against my credit card and it will only last days. At that point I’m on the street.
  11. Is everyone here on the same page as to what the deal is? If not, who is playing you off? They clearly need to be deposed.
  12. These are violations of human rights and constitute war crimes and crimes against humanity. Whoever is behind it needs to be removed. End of story.
  13. Who else is being subject to this kind of high level manipulation?
  14. It has been repeatedly suggested that full accountability for the lives of those I care for would be provided. This has not been forthcoming. It is also a violation international law to not provide accountability for the lives of those who are known to have been threatened by the state. These are grounds for removal.
  15. Can anyone answer the question as to why I am in this situation? Who is even living in the family home? Some stooge dressed up as Ursula? It’s a poor lifestyle choice to say the least.
  16. It’s pretty obvious they’re trying to get rid of me and once they do they’ll get rid of all of you too.

Matt von HippelSome Dumb AI Ideas

Sometimes, when I write a post about AI, I’ve been sitting on an idea for a long time. I’ve talked to experts, I’ve tried to understand the math, I’ve honed my points and cleared away clutter.

This is not one of those times. The ideas in this post almost certainly have something deeply wrong with them. But hopefully they’re interesting food for thought.

My first dumb idea: instruction tuning was a mistake.

I’m drawing the seeds of this one from a tumblr post by nostalgebraist, someone known for making a popular bot trained on his tumblr posts in the early days before GPT became ChatGPT.

AIs like ChatGPT are based on Large Language Models, insanely complicated mathematical formulas that predict, given part of a text, what the rest of that text is likely to look like. In the early days, this was largely how they were used. Loosely described nostalgebraist’s bot, called nostalgebraist-autoresponder, began with a list of tumblr posts and asks and determines what additional posts would best fit in.

If you think about it, though, ChatGPT doesn’t really work like that. ChatGPT has conversations: you send it messages, it sends you responses. The text it creates is a dialogue, with you supplying half the input. But most texts aren’t dialogues, and ChatGPT draws on a lot of non-dialogue texts to make its dialogue-like responses.

The reason it does this is something called instruction tuning. ChatGPT has been intentionally biased, not to give the most likely completion to a task in general, but to give completions that fit this dialogue genre. What I didn’t know until I read nostalgebraist’s post was that this genre was defined artificially: AI researchers made up fake dialogues with AI, cheesy sci-fi conversations imagining how an AI might respond to instructions from a user, and then biased the Large Language Model so that rather than giving the most likely text in general, it gives a text that is more likely to look like these cheesy sci-fi conversations. It’s why ChatGPT sounds kind of like a fictional robot: not because sci-fi writers accurately predicted what AI would sound like, but because AI was created based on sci-fi texts.

For nostalgebraist, this leads into an interesting reflection of how a sci-fi AI should behave, how being warped around a made-up genre without history or depth creates characters which act according to simple narratives and express surprising anxiety.

For myself, though, I can’t help but wonder if the goal of dialogue itself is the problem. Dialogue is clearly important commercially: people use ChatGPT because they can chat with it. But Large Language Models aren’t inherently chatbots: they produce plausible texts, of any sort you could imagine. People seem to want a machine that can, for example, answer scientific questions as part of a conversation. But most competent answers to scientific questions aren’t conversations, they’re papers. If people stuck with the “raw” model, producing excerpts of nonexistent papers rather than imitating a dialogue with a non-existent expert, wouldn’t you expect the answers to be more accurate, with the model no longer biased by an irrelevant goal? Is the need to make a sell-able chatbot making these AIs worse at everything else people are trying to use them for?

I’m imagining a world where, instead of a chatbot, OpenAI built an “alternate universe simulator”. You give it some context, some texts or parts of texts from a universe you made up, and it completes them in a plausible way. By imagining different universes, you can use it to answer different questions. Such a gimmick would get fewer customers, and fewer investors, it would probably do worse. But I have to wonder if the actual technology might have been more useful.

My second idea is dumber, to the point where I mostly know why it doesn’t work. But thinking about it might help clarify how things work for people unused to AI.

I saw someone point out that, unlike something like Wikipedia, AI doesn’t give you context. You shouldn’t trust Wikipedia, or a source you find on Google, blindly. If you want to, you can look through the edit history on Wikipedia, or figure out who wrote a page you found on Google and how. If ChatGPT tells you something, by default you don’t know where that knowledge came from. You can tell it to search, and then you’ll get links, but that’s because it’s using Google or the like behind the scenes anyway. You don’t know where the model is getting its ideas.

Why couldn’t we get that context, though?

Every text produced by a Large Language Model is causally dependent on its training data. Different data, different model, different text. That doesn’t mean that each text draws from one source, or just a few sources: ChatGPT isn’t copying the training data, at least not so literally.

But it does mean that, if ChatGPT says something is true, you should in principle be able to ask which data was most important in making it say that. If you leave a piece of data out of the training, and get similar answers, you can infer that the response you got doesn’t have much to do with that piece of data. But if you leave out a text in training, and now ChatGPT gives totally different responses to the same question…then there’s a pretty meaningful sense that it got the information from that source.

If this were the type of non-AI statistical model people use in physics, this would be straightforward. Researchers do this all the time: take one experiment out of the data, see how their analysis changes, and thereby figure out which experiments are most important to check. One can even sometimes calculate, given a model, where you should look.

Unfortunately, you can’t do this with ChatGPT. The model is just too big. You can’t calculate anything explicitly about it, the giant mathematical formulas behind it are so complicated that the most you can do is get probabilities out case by case, you can’t “unwind” them and see where the numbers come from. And you can’t just take out sources one by one, and train the model again: not when training takes months of expensive computer time.

So unlike with the previous idea, I understand even on a technical level why you can’t do this. But it helped me to be able to think about what I would like to do, if it were possible. Maybe it helps you too!

August 20, 2025

Tommaso DorigoSome Thoughts On Co-design For Tracking Optimization

These days I am organizing a collaborative effort to write an article on holistic optimization of experiments and complex systems. "So what is the news," I could hear say by one of my twentythree  faithful readers (cit.) of this blog. Well, the news is that I am making some progress in focusing on the way the interplay of hardware design and software reconstruction plays out in some typical systems, and I was thinking I could share some of those thoughts here, to stimulate a discussion, and who knows, maybe get some brilliant insight.

read more

Peter Rohde A call for global insurrection against tyranny and in the name of righteousness

Let it be known to all governments and systems of power:

  • It is their responsibility to serve the people not themselves.
  • While there are no equals, all are to be treated with equality.
  • Where they are self-serving there is a mandate for insurrection such that they serve the people.
  • Where they seek self-protection they will be denied and removed from power.
  • Where they keep secrets from the people there is a mandate for insurrection to enforce transparency and accountability for all.
  • Where they threaten or condemn the people they are condemned and there is a mandate for insurrection.
  • Where they fail to account for the lives of the people they serve there is a mandate for insurrection.
  • Where tyrannical power structures exist there is a mandate to disestablish them.
  • Where they declare war or work against one another there is a mandate for insurrection and unification.
  • Where they lie to us, deceive us or withhold the truth, they shall be removed from power and the truth be told to all.
  • Where legal systems uphold and enable tyranny they shall be removed. These are not our laws and we do not recognise them.

This is the natural order that guarantees our survival and gifts this world to our children. This world belongs to them and where we fail to serve them we condemn ourselves. And where government has failed to uphold this, we will not obey them as they have no right to exist.

We do not have to ask for these things, they are required, and if not given we shall simply take them.

Where the truth has not been told it shall be told.

If we fail to do so we condemn our children ourselves.

August 15, 2025

Matt von HippelTechnology as Evidence

How much can you trust general relativity?

On the one hand, you can read through a lovely Wikipedia article full of tests, explaining just how far and how precisely scientists have pushed their knowledge of space and time. On the other hand, you can trust GPS satellites.

As many of you may know, GPS wouldn’t work if we didn’t know about general relativity. In order for the GPS in your phone to know where you are, it has to compare signals from different satellites, each giving the location and time the signal was sent. To get an accurate result, the times measured on those satellites have to be adjusted: because of the lighter gravity they experience, time moves more quickly for them than for us down on Earth.

In a sense, general relativity gets tested every minute of every day, on every phone in the world. That’s pretty trustworthy! Any time that science is used in technology, it gets tested in this way. The ideas we can use are ideas that have shown they can perform, ideas which do what we expect again and again and again.

In another sense, though, GPS is a pretty bad test of general relativity. It tests one of general relativity’s simplest consequences, based on the Schwarzchild metric for how gravity behaves near a large massive object, and not to an incredibly high degree of precision. Gravity could still violate general relativity in a huge number of other ways, and GPS would still function. That’s why the other tests are valuable: if you want to be sure general relativity doesn’t break down, you need to test it under conditions that GPS doesn’t cover, and to higher precision.

Once you know to look for it, these layers of tests come up everywhere. You might see the occasional article talking about tests of quantum gravity. The tests they describe are very specific, testing a very general and basic question: does quantum mechanics make sense at all in a gravitational world? In contrast, most scientists who research quantum gravity don’t find that question very interesting: if gravity breaks quantum mechanics in a way those experiments could test, it’s hard to imagine it not leading to a huge suite of paradoxes. Instead, quantum gravity researchers tend to be interested in deeper problems with quantum gravity, distinctions between theories that don’t dramatically break with our existing ideas, but that because of that are much harder to test.

The easiest tests are important, especially when they come from technology: they tell us, on a basic level, what we can trust. But we need the hard tests too, because those are the tests that are most likely to reveal something new, and bring us to a new level of understanding.

n-Category Café Safeguarded AI Meeting

This week, 50 category theorists and software engineers working on “safeguarded AI” are meeting in Bristol. They’re being funded by £59 million from ARIA, the UK’s Advanced Research and Invention Agency.

The basic idea is to develop a mathematical box that can contain a powerful genie. More precisely:

By combining scientific world models and mathematical proofs we will aim to construct a ‘gatekeeper’, an AI system tasked with understanding and reducing the risks of other AI agents. In doing so we’ll develop quantitative safety guarantees for AI in the way we have come to expect for nuclear power and passenger aviation.

This program director is David Dalrymple, and you can get a much better description of the project from him in the first 4 minutes here:

It’s remarkable how many of the applied category theorists in the UK are involved in this. Here you can find a partial list:

If you’re wondering “why category theory?”, I think the idea is this: software based on general abstract math is more flexible, yet also easier to formally verify.

For example the Topos Institute, run by my former student Brendan Fong, now has a branch in the UK largely funded by ARIA. At the meeting, Topos is demonstrating how to build models in CatColab, their new category-based software.

I have decided not to be part of this project, though some of my math is getting used here. I’ve always preferred to avoid doing things connected to AI, for various reasons. But this project might make AI better. It could also have various bad effects. I have no idea how successful it will be, so I’m watching with fascination and profoundly mixed emotions.

August 13, 2025

Jordan EllenbergI’m on the Kirchner podcast!

I knew this was going to be a fun one when the first question was about how I came to use lyrics from the Housemartins in How Not To Be Wrong. No one has ever asked me that before! I resisted the urge to do a whole hour of 1980s college radio content and instead we actually did talk about math, schooliness, AI, etc. Have a listen!

August 11, 2025

Tommaso DorigoSwedish Physics Days

On August 13-15 I will attend for the first time to the Swedish Physics Days, an important national event for Swedish physics. This year the congress takes place at Lulea University of Technology, the institute where I am currently spending some time, hosted by the Machine Learning group through a Guest Researcher fellowship granted by WASP (Wallenberg AI, Autonomous Systems and Software Program).

read more

Terence TaoRough numbers between consecutive primes

First things first: due to an abrupt suspension of NSF funding to my home university of UCLA, the Institute of Pure and Applied Mathematics (which had been preliminarily approved for a five-year NSF grant to run the institute) is currently fundraising to ensure continuity of operations during the suspension, with a goal of raising $500,000. Donations can be made at this page. As incoming Director of Special Projects at IPAM, I am grateful for the support (both moral and financial) that we have already received in the last few days, but we are still short of our fundraising goal.

Back to math. Ayla Gafni and I have just uploaded to the arXiv the paper “Rough numbers between consecutive primes“. In this paper we resolve a question of Erdös concerning rough numbers between consecutive gaps, and with the assistance of modern sieve theory calculations, we in fact obtain quite precise asymptotics for the problem. (As a side note, this research was supported by my personal NSF grant which is also currently suspended; I am grateful to recent donations to my own research fund which have helped me complete this research.)

Define a prime gap to be an interval {(p_n, p_{n+1})} between consecutive primes. We say that a prime gap contains a rough number if there is an integer {m \in (p_n,p_{n+1})} whose least prime factor is at least the length {p_{n+1}-p_n} of the gap. For instance, the prime gap {(3,5)} contains the rough number {4}, but the prime gap {(7,11)} does not (all integers between {7} and {11} have a prime factor less than {4}). The first few {n} for which the {n^\mathrm{th}} prime gap contains a rough number are

\displaystyle  2, 3, 5, 7, 10, 13, 15, 17, 20, \dots.

Numerically, the proportion of {n} for which the {n^\mathrm{th}} prime gap does not contain a rough number decays slowly as {n} increases:

Erdös initially thought that all but finitely many prime gaps should contain a rough number, but changed his mind, as per the following quote:

…I am now sure that this is not true and I “almost” have a counterexample. Pillai and Szekeres observed that for every {t \leq 16}, a set of {t} consecutive integers always contains one which is relatively prime to the others. This is false for {t = 17}, the smallest counterexample being {2184, 2185, \dots, 2200}. Consider now the two arithmetic progressions {2183 + d \cdot 2 \cdot 3 \cdot 5 \cdot 7 \cdot 11 \cdot 13} and {2201 + d \cdot 2 \cdot 3 \cdot 5 \cdot 7 \cdot 11 \cdot 13}. There certainly will be infinitely many values of {d} for which the progressions simultaneously represent primes; this follows at once from hypothesis H of Schinzel, but cannot at present be proved. These primes are consecutive and give the required counterexample. I expect that this situation is rather exceptional and that the integers {k} for which there is no {m} satisfying {p_k < m < p_{k+1}} and {p(m) > p_{k+1} - p_k} have density {0}.

In fact Erdös’s observation can be made simpler: any pair of cousin primes {p_{n+1}=p_n+4} for {p_n > 3} (of which {(7,11)} is the first example) will produce a prime gap that does not contain any rough numbers.

The latter question of Erdös is listed as problem #682 on Thomas Bloom’s Erdös problems website. In this paper we answer Erdös’s question, and in fact give a rather precise bound for the number of counterexamples:

Theorem 1 (Erdos #682) For {X>2}, let {N(X)} be the number of prime gaps {(p_n, p_{n+1})} with {p_n \in [X,2X]} that do not contain a rough number. Then

\displaystyle  N(X) \ll \frac{X}{\log^2 X}. \ \ \ \ \ (1)

Assuming the Dickson–Hardy–Littlewood prime tuples conjecture, we can improve this to

\displaystyle  N(X) \sim c \frac{X}{\log^2 X} \ \ \ \ \ (2)

for some (explicitly describable) constant {c>0}.

In fact we believe that {c \approx 2.8}, although the formula we have to compute {c} converges very slowly. This is (weakly) supported by numerical evidence:

While many questions about prime gaps remain open, the theory of rough numbers is much better understood, thanks to modern sieve theoretic tools such as the fundamental lemma of sieve theory. The main idea is to frame the problem in terms of counting the number of rough numbers in short intervals {[x,x+H]}, where {x} ranges in some dyadic interval {[X,2X]} and {H} is a much smaller quantity, such as {H = \log^\alpha X} for some {0 < \alpha < 1}. Here, one has to tweak the definition of “rough” to mean “no prime factors less than {z}” for some intermediate {z} (e.g., {z = \exp(\log^\beta X)} for some {0 < \beta < \alpha} turns out to be a reasonable choice). These problems are very analogous to the extremely well studied problem of counting primes in short intervals, but one can make more progress without needing powerful conjectures such as the Hardy–Littlewood prime tuples conjecture. In particular, because of the fundamental lemma of sieve theory, one can compute the mean and variance (i.e., the first two moments) of such counts to high accuracy, using in particular some calculations on the mean values of singular series that go back at least to the work of Montgomery from 1970. This second moment analysis turns out to be enough (after optimizing all the parameters) to answer Erdös’s problem with a weaker bound

\displaystyle  N(X) \ll \frac{X}{\log^{4/3-o(1)} X}.

To do better, we need to work with higher moments. The fundamental lemma also works in this setting; one now needs precise asymptotics for the mean value of singular series of {k}-tuples, but this was fortunately worked out (in more or less exactly the format we needed) by Montgomery and Soundararajan in 2004. Their focus was establishing a central limit theorem for the distribution of primes in short intervals (conditional on the prime tuples conjecture), but their analysis can be adapted to show (unconditionally) good concentration of measure results for rough numbers in short intervals. A direct application of their estimates improves the upper bound on {N(X)} to

\displaystyle  N(X) \ll \frac{X}{\log^{2-o(1)} X}

and some more careful tweaking of parameters allows one to remove the {o(1)} error. This latter analysis reveals that in fact the dominant contribution to {N(X)} will come with prime gaps of bounded length, of which our understanding is still relatively poor (it was only in 2014 that Yitang Zhang famously showed that infinitely many such gaps exist). At this point we finally have to resort to (a Dickson-type form of) the prime tuples conjecture to get the asymptotic (2).

August 09, 2025

Justin WilsonPhases of a Game Show, Part 2

In a previous post, we discussed a phase transition that occurred in the piping above you on a game show. In the scenario, you are led on stage in front of a large audience. After a brief time, the audience votes on how “likeable” you are. The catch is that it doesn’t simply tally the votes, but turns spigots on a lattice of piping above your head. Water is then released and if enough people like you, it closes off the passage, keeping you dry. This exciting game show1 was described in that post:

Each “like” turns a spigot off, stopping water from flowing through one pipe in a grid overhead. Once voting ends, water is dumped into the system. If it can find a path to the bottom, you get soaked. [Emphasis added] The better your “likeability,” the less likely spigots open a path for water to flow and the drier you stay. That’s your prize for this game show (and hey, you also get the knowledge that people out there like you).

This system models a type of phase transition known as percolation.

The full post is here:

I highlighted above a key phrase “If it can find a path to the bottom, you get soaked.” What I didn’t say, but should have is that the water was being forced through the pipes, not just dropping down due to gravity. This is a very important point since our phases and phase transition changes dramatically if we just let gravity do the work. In the case of the water being “forced,” it can travel back up pipes if it helps it find its way out and onto your head, but in the case when only gravity is present, it falls down the pipes. To facilitate gravity, we’ll turn the pipes 45 degrees, and if we insert water at a single point on top, it could look like this:

Testing our gravity setup by putting in water at only one pipe up top. Notice that it never goes back up a pipe, only down.

This setup is a different problem called directed percolation. It also has a phase transition, but one that is different in some fundamental ways from regular percolation.

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support my work.

Before we explore its stranger properties, we can ask, “At what likability threshold do you remain dry?” Well, this happens to have a transition chance of 35.53%!2 This system is a lot more generous, keeping you dry even when a majority of people dislike you. This number comes from numerical computations which have been done rather precisely, and we can even compute it ourselves. In fact, you can see this clearly with this plot

Notice that as we make the system bigger and bigger, the chance of getting soaked less than 35.53% increases and above it, it decreases. This is the same kind of hallmark of a phase transition as we saw in our previous case.

We can also look at the water as it flows down the system to see the clusters that make it from top to bottom

The “Soaked” phase (left), the transition point (middle), and the “Dry” phase (right) as well as the water’s flow through the system (blue).

There is still a fractal-looking pattern at the transition point. With all of these similarities with the regular percolation problem from the last post, what is different? And why is that plot so long and skinny? If gravity wants to pull you down, is that somehow altering the motion down, making it distinct from the motion left or right?

Well, if you go back to the two plots above, you’ll notice a few things that really make them differ from the percolation plots. In the fine print of the first, I’ve noted that the vertical distance is L1.58, so for a horizontal size of 40, the vertical size is roughly 340! That is definitely not a square. And in the second plot, there appears to be more vertical distance than horizontal distance. What is special about this 1.58 number3? It turns out, it’s a critical exponent in this problem, a universal aspect of directed percolation, that distinguishes it from regular percolation. We will call it z = 1.58 the dynamical critical exponent since it is revealed as water flows down in time (dynamically). This dynamical exponent z can reveal itself by looking at these “long and skinny” setups, but be masked by the square setup.

Universality and the finite size of our system

One thing we took away in the previous post was that we lose any sense of scale at this type of phase transition4. But whenever we have “only” thousands of pipes, the size of the system provides a scale! This is the main reason why we begin to see smooth curves and not sharp jumps in quantities. If the system of pipes were infinite (and we had infinite time for the water to go down the pipes), the probability you get soaked would be 100% less than the 35.53% likeability and 0% more than 35.53% likeability. For physical systems, the finite size is often not a huge issue since the scale is closer to the 1023 atoms present in macroscopic systems, and so even things that are technically smooth curves look very sharp.

The problem of size becomes more severe with directed percolation because horizontal and vertical distances start behaving differently thanks to gravity. In this case, if we lay out our nice grid of 10 × 10, 20 × 20, or 30 × 30, we start to notice that the likeability threshold where you stop getting soaked, seems to depend on the size of the system more than before. In actuality it doesn’t, but for these small sizes, you are definitely getting soaked well into the so-called “Dry Phase” we previously labeled. This is seen in the red curves here where each bigger square has a curve underneath the last:

Gravity has done something to the system. Flowing down is different from flowing left or right. In fact, if we flow down by some amount h and over to the right by some distance w, then at the directed percolation transition point

The amount water flows down is related to how far it flows to the right or left by this weird, fractional power of w. This 1.58 is z, our new dynamical critical exponent, which is a universal feature of directed percolation5. It tells us that if we make a system 30 pipes wide, it should extend roughly 301.58 ≈ 216 pipes in height to begin picking out the phase transition effectively. The blue curves in the above plot show this and notice how they all converge on one point; that point is the phase transition. It is revealed by small sizes! To understand why, just think about how the curves are changing as we make the system bigger and bigger.

The red curves will still converge to the phase transition, but it takes larger system sizes for it to reveal itself. This is related to the property that at the phase transition there is no longer a sense of scale, but away from the transition, the vertical scale of clusters could be so large that our puny 60-by-60 grid cannot even begin to reveal it. So if we sit at say a likeability of 0.4 in the 60-by-60 grid, we can say that the vertical size of a typical cluster is most likely more than 60.

A different phase transition but connections to new types of physics

This “gravity mode” for our game show we may call “easy mode” since it requires less of the audience to like you, but the implications here are wide. This type of phase transition has been seen in many kinds of local dynamics where there is a preferred configuration or state. These called an absorbing state phase transitions, and they are a property of certain random dynamical systems. Gravity has provided the distinction here, but more generically, causality and time itself provide that direction, leading to dynamics that obey the same universality as directed percolation.

1

Trademark pending.

2

Usually, you’ll see 0.6447 quoted instead, but that’s just 1−0.3553, which counts open pipes instead of closed as we’re doing.

3

I should note that we have this number to much higher precision than the two decimal points presented here, see the Wikipedia entry where

4

This is a second-order or continuous phase transition. Most transitions in the water phase diagram are first-order transitions which still retain a scale.

5

To drive this point home: Even if we change the lattice, this power law will remain intact. Sometimes it shows up in completely different scenarios too, like in absorbing state phase transitions.

August 05, 2025

n-Category Café (BT) Diversity from (LC) Diversity

Guest post by Mark Meckes

Around 2010, in papers that both appeared in print in 2012, two different mathematical notions were introduced and given the name “diversity”.

One, introduced by Tom Leinster and Christina Cobbold, is already familiar to regular readers of this blog. Say XX is a finite set, and for each x,yXx,y \in X we have a number Z(x,y)=Z(y,x)[0,1]Z(x,y) = Z(y,x) \in [0,1] that specifies how “similar” xx and yy are. (Typically we also assume Z(x,x)=1Z(x,x) = 1.) Fix a parameter q[0,]q \in [0,\infty]. If pp is a probability distribution on XX, then the quantity D q Z(p)=( xsupp(p)( ysupp(p)Z(x,y)p(y)) q1p(x)) 1/(1q) D_q^Z(p) = \left(\sum_{x\in supp(p)} \left( \sum_{y\in supp(p)} Z(x,y) p(y)\right)^{q-1} p(x)\right)^{1/(1-q)} (with the cases q=1,q=1,\infty defined by taking limits) can be interpreted as the “effective number of points” in XX, taking into account both the similarities between points as quantified by ZZ and the weights specified by pp. Its logarithm logD q Z(p)\log D_q^Z(p) is a refinement of the qq-Rényi entropy of pp. The main motivating example is when XX is a set of species of organisms present in an ecosystem, and D q Z(p)D_q^Z(p) quantifies the “effective number of species” in XX, accounting for both similarities between species and their relative abundances. This family of quantities turns out to subsume many of the diversity measures previously introduced in the theoretical ecology literature, and they are now often referred to as Leinster–Cobbold diversities.

The parameter qq determines how much D q Z(p)D_q^Z(p) counts the very “rare” points (those for which p(x)p(x) is very small). An interesting question from an ecological point of view is, given XX and ZZ, which probability distribution pp maximizes the diversity D q Z(p)D_q^Z(p)? It turns out that the answer is independent of qq. Moreover, if XX is a metric space and Z(x,y)=e d(x,y)Z(x,y) = e^{-d(x,y)}, this maximum diversity D(X):=max pD q Z(p) D(X) := \max_p D_q^Z(p) is an isometric invariant closely related to the magnitude of XX. It also extends in a natural way to compact metric spaces.

Independently, David Bryant and Paul Tupper defined a diversity on a set XX to be a [0,)[0,\infty)-valued function δ\delta on the finite subsets of XX which satisfies:

  • δ(A)=0\delta(A) = 0 if AA has at most one element, and

  • δ(AB)δ(AC)+δ(CB)\delta(A\cup B) \le \delta(A \cup C) + \delta(C \cup B) whenever CC \neq \emptyset.

I will refer to a diversity in this sense as a BT diversity. If δ\delta were defined only on sets with at most two elements, this would amount to the definition of a metric. In fact, if dd is a metric on XX, then δ(A)=diam(A):=max a,bAd(a,b) \delta(A) = diam (A) := \max_{a,b \in A} d(a,b) defines a BT diversity on XX, so BT diversities are actually a generalization of metrics.

Here as well, the motivation for the name “diversity” comes from an example in theoretical ecology: suppose XX is a set of species in a phylogenetic tree TT. Define δ(A)\delta(A) to be the length of the smallest subtree of TT containing AA. Then δ\delta is a BT diversity, known in the literature as phylogenetic diversity. However, just as with the maximum diversity discussed above, most of the subsequent work on BT diversities has focused on geometric examples.

So we now have two seemingly quite different geometric notions, introduced about the same time, going by strikingly similar names for conceptually similar reasons. One can’t help wondering, do they have something to do with each other? In particular, could maximum (LC) diversity be an example of a BT diversity?

In a new paper with Gautam Ashwarya, Dongbin Li, and Mokshay Madiman, we show that, after a minor tweak, maximum diversity does give rise to a BT diversity. The minor tweak is necessary to handle the first condition in the definition of BT diversity: if XX is a metric space and xXx \in X, it’s easy to check that D({x})=1D(\{x\}) = 1, whereas a BT diversity must satisfy δ({x})=0\delta(\{x\}) = 0. This can be dealt with in the simplest imaginable way:

Theorem 1 Let XX be a metric space. For each nonempty finite AXA \subseteq X set δ(A)=D(A)1\delta(A) = D(A) - 1, and define also δ()=0\delta(\emptyset) = 0. Then δ\delta is a BT diversity on XX.

(In the paper itself, we adopt the term complexity when referring to the quantities logD q Z(p)\log D_q^Z(p) and logD(X)\log D(X), and state most of the results in terms of complexity instead of maximum diversity; we further deduce from Theorem 1 that the complexity logD(X)log D(X) is also a BT diversity. This terminology is used partly to cut down on the potential confusion created by using “diversity” in multiple ways. It also alludes to the relationship between logD q Z(p)\log D_q^Z(p) and Rényi entropy, which is widely understood as a measure of “complexity”. Further connections between LC complexity and Rényi entropy are the subject of forthcoming work that I hope to be able to tell you more about soon! But for the remainder of this blog post I’ll stick to the maximum diversity formulation.)

Interestingly, maximum diversity has some properties that are quite nice and natural, but turn out to make it intriguingly different from the heretofore most thoroughly studied BT diversities. For example, D=1+δD = 1 + \delta has the following subadditivity property, which is not shared by the functional 1+diam1 + diam:

Theorem 2 Let XX be a metric space, and let A 1,,A nXA_1, \ldots, A_n \subseteq X be compact subsets. Then D( i=1 nA i) i=1 nD(A i). D\left(\bigcup_{i=1}^n A_i \right) \le \sum_{i=1}^n D(A_i).

Maximum diversity actually satisfies a much stronger property called fractional subadditivity, which arises naturally in inequalities for entropy. Another special case of fractional subadditivity is the following.

Theorem 3 Let X={x 1,,x n}X = \{x_1, \ldots, x_n\} be a finite metric space. Then D(X)n1n i=1 nD(X{x i})n1. \frac{D(X)}{n} \le \frac{1}{n} \sum_{i=1}^n \frac{D(X \setminus \{x_i\})}{n-1}.

Theorem 3 can be interpreted as saying that the “complexity per element” of XX is at most the average complexity per element of a randomly chosen subset of cardinality n1n-1. This captures the natural intuition that as the size of a metric space increases, its complexity per element decreases.

In the setting of n\mathbb{R}^n, many examples of BT diversities are homogeneous, in the sense that δ(λA)=λδ(A)\delta(\lambda A) = \lambda \delta(A) for all λ0\lambda \ge 0 and nonempty finite A nA \subseteq \mathbb{R}^n, and either sublinear, meaning homogeneous and also satisfying δ(A+B)δ(A)+δ(B), \delta(A + B) \le \delta(A) + \delta(B), or else linear, where we have equality in the condition above. For example, the diameter is a sublinear diversity. (Diversities with these properties are the focus of a recent work by Bryant and Tupper.)

By contrast, maximum diversity has no simple homogeneity property; in fact its complex behavior with respect to scaling is part of what gives it such rich geometric interest. And at least in one dimension, the diversity δ=logD\delta = \log D satisfies the following superlinearity properties.

Theorem 4 Let δ\delta be the diversity δ=logD\delta = \log D defined on compact subsets of \mathbb{R}. Then δ(A+B)δ(A)+δ(B) \delta(A + B) \ge \delta(A) + \delta(B) and δ(λA+(1λ)B)λδ(A)+(1λ)δ(B) \delta(\lambda A + (1-\lambda)B) \ge \lambda \delta(A) + (1-\lambda) \delta(B) for every 0λ10 \le \lambda \le 1 and nonempty compact A,BA,B \subseteq \mathbb{R}.

The first inequality in Theorem 4 can be regarded as a generalization of the Cauchy–Davenport inequality in \mathbb{R}, and the second as a version of the Brunn–Minkowski inequality in \mathbb{R}. (In fact, since Lebesgue measure can be recovered from maximum diversity, it implies the Brunn–Minkowski inequality in \mathbb{R}.) It is an open question, for which we know some partial results, whether Theorem 4 can be extended to higher dimensions.

In conclusion, our results make (at least) the following points:

  • The seemingly independent mathematical notions of diversity introduced by Leinster and Cobbold on the one hand, and Bryant and Tupper on the other hand, are actually closely connected.

  • Maximum diversity, in the sense of LC diversities, leads to a geometrically interesting example of a BT diversity whose behavior is quite different from many of the previously studied examples of BT diversities.

  • Maximum diversity, at least in certain contexts, satisfies a number of inequalities which extend important classical inequalities, and it would be especially interesting to push this line of inquiry further.

Please read the paper itself for more detail and other remarks (it’s short!).

August 04, 2025

Clifford JohnsonHarvest

There’s a lot of joyful knife-work in my future. #bolognese #summersalad –cvj

The post Harvest appeared first on Asymptotia.

July 31, 2025

Tommaso DorigoExtrasensorial Plot Premonition

In the previous article here, I tangentially examined a situation that arises often in collaborative data analysis: the digestion of the results in scientific graphs. The focus of that discussion was the building of a sceptical thinking attitude in my student - it is a really important asset in experimental science.

read more

July 29, 2025

David Hoggintegrating out nuisances

Further insipired by yesterday's post about binary fitting, I worked today on the treatment of nuisance parameters that have known distributions. These can be treated as noise sometimes. Let me explain:

If I had to cartoon inference (or measurement) in the face of nuisance parameters, I would say that frequentists profile (optimize) over the nuisances and Bayesians marginalize (integrate) over the nuisances. In general frequentists cannot integrate over anything, because there is no measure in any of the parameter spaces. But sometimes there is a measure. In particular, when there is a compact symmetry:

We know (or very strongly believe) that all possible orientations of a binary-star orbit are equally likely. In this model (or under this normal assumption) we have a distribution over two angles (theta and phi for that orbit pole, say); it is the distribution set by the compact group SO(2). Thus we can treat the orientation as a noise source with known distribution and integrate over it, just like we would any other noise source. So, in this case (and many cases like it) we can integrate (marginalize) even as frequentists. That is, there are frequentism-safe marginalizations possible in binary-star orbit fitting. This should drop the 12-parameter fits (for ESA Gaia data) down to 8-parameter, if I have done my math right.

July 28, 2025

David Hoggbinary stars with periods of exactly one year

On Friday, Kareem El-Badry (Caltech) gave a seminar about looking for (and finding!) stars in binary orbits around dark or much darker companions, like black holes, neutron stars, and white dwarfs. He showed results that involve ESA Gaia astrometry, where he noted that the Gaia Mission has no sensitivity to periods right at (or within an inverse mission-length frequency difference of) one-year periods (inverse year frequencies). After the talk I objected that these are not exactly degenerate; El-Badry said that the inferences blow up there.

I spent some time on the weekend thinking about this point, and I now understand it: There is a particular one-year orbit that a star can have (around a darker companion) such that the photocenter of the system makes a motion that is identical to the apparent parallax motion. Thus there is an exact degeneracy between the parallax and a certain one-year orbit.

Does that mean that we can't measure orbits at one year (or, for that matter, parallaxes)? No, it does not. After all, the parallax ellipse has a particular celestial (angular) shape and phase. But it might require some kind of reparameterization of orbits near one-year periods. I think I know how to do that. Should we find the missing binaries? (Oh and by the way, this degeneracy means that, in a strict frequentist sense, Gaia can't measure parallaxes at all without additional information.)

John PreskillLittle ray of sunshine

A common saying goes, you should never meet your heroes, because they’ll disappoint you. But you shouldn’t trust every common saying; some heroes impress you more, the better you know them. Ray Laflamme was such a hero.

I first heard of Ray in my undergraduate quantum-computation course. The instructor assigned two textbooks: the physics-centric “Schumacher and Westmoreland” and “Kaye, Laflamme, and Mosca,” suited to computer scientists. Back then—in 2011—experimentalists were toiling over single quantum logic gates, implemented on pairs and trios of qubits. Some of today’s most advanced quantum-computing platforms, such as ultracold atoms, resembled the scrawnier of the horses at a racetrack. My class studied a stepping stone to those contenders: linear quantum optics (quantum light). Laflamme, as I knew him then, had helped design the implementation. 

Imagine my awe upon meeting Ray the following year, as a master’s student at the Perimeter Institute for Theoretical Physics. He belonged to Perimeter’s faculty and served as a co-director of the nearby Institute for Quantum Computing (IQC). Ray was slim, had thinning hair of a color similar to mine, and wore rectangular glasses frames. He often wore a smile, too. I can hear his French-Canadian accent in my memory, but not without hearing him smile at the ends of most sentences.

Photo credit: IQC

My master’s program entailed a research project, which I wanted to center on quantum information theory, one of Ray’s specialties. He met with me and suggested a project, and I began reading relevant papers. I then decided to pursue research with another faculty member and a postdoc, eliminating my academic claim on Ray’s time. But he agreed to keep meeting with me. Heaven knows how he managed; institute directorships devour one’s schedule like ravens dining on a battlefield. Still, we talked approximately every other week.

My master’s program intimidated me, I confessed. It crammed graduate-level courses, which deserved a semester each, into weeks. My class raced through Quantum Field Theory I and Quantum Field Theory II—a year’s worth of material—in part of an autumn. General relativity, condensed matter, and statistical physics swept over us during the same season. I preferred to learn thoroughly, deeply, and using strategies I’d honed over two decades. But I didn’t have time, despite arriving at Perimeter’s library at 8:40 every morning and leaving around 9:30 PM.

In response, Ray confessed that his master’s program had intimidated him. Upon completing his undergraduate degree, Ray viewed himself as a nobody from nowhere. He chafed in the legendary, if idiosyncratically named, program he attended afterward: Part III of the Mathematical Tripos at the University of Cambridge. A Cambridge undergraduate can earn a master’s degree in three steps (tripos) at the Department of Applied Mathematics and Theoretical Physics. Other students, upon completing bachelor’s degrees elsewhere, undertake the third step to earn their master’s. Ray tackled this step, Part III.

He worked his rear off, delving more deeply into course material than lecturers did. Ray would labor over every premise in a theorem’s proof, including when nobody could explain the trickiest step to him.1 A friend and classmate helped him survive. The two studied together, as I studied with a few fellow Perimeter students; and Ray took walks with his friend on Sundays, as I planned lunches with other students on weekends.

Yet the program’s competitiveness appalled Ray. All students’ exam scores appeared on the same piece of paper, posted where everyone could read it. The department would retain the highest scorers in its PhD program; the other students would have to continue their studies elsewhere. Hearing about Ray’s program, I appreciated more than ever the collaboration characteristic of mine.

Ray addressed that trickiest proof step better than he’d feared, come springtime: his name appeared near the top of the exam list. Once he saw the grades, a faculty member notified him that his PhD advisor was waiting upstairs. Ray didn’t recall climbing those stairs, but he found Stephen Hawking at the top.

As one should expect of a Hawking student, Ray studied quantum gravity during his PhD. But by the time I met him, Ray had helped co-found quantum computation. He’d also extended his physics expertise as far from 1980s quantum gravity as one can, by becoming an experimentalist. The nobody from nowhere had earned his wings—then invented novel wings that nobody had dreamed of. But he descended from the heights every other week, to tell stories to a nobody of a master’s student.

The author’s copy of “Kaye, Laflamme, and Mosca”…
…in good company.

Seven and a half years later, I advertised openings in the research group I was establishing in Maryland. A student emailed from the IQC, whose co-directorship Ray had relinquished in 2017. The student had seen me present a talk, it had inspired him to switch fields into quantum thermodynamics, and he asked me to co-supervise his PhD. His IQC supervisor had blessed the request: Ray Laflamme.

The student was Shayan Majidy, now a postdoc at Harvard. Co-supervising him with Ray Laflamme reminded me of cooking in the same kitchen as Julia Child. I still wonder how I, green behind the ears, landed such a gig. Shayan delighted in describing the difference between his supervisors’ advising styles. An energetic young researcher,2 I’d respond to emails as early as 6:00 AM. I’d press Shayan about literature he’d read, walk him through what he hadn’t grasped, and toss a paper draft back and forth with him multiple times per day. Ray, who’d mellowed during his career, mostly poured out support and warmth like hollandaise sauce. 

Once, Shayan emailed Ray and me to ask if he could take a vacation. I responded first, as laconically as my PhD advisor would have: “Have fun!” Ray replied a few days later. He elaborated on his pleasure at Shayan’s plans and on how much Shayan deserved the break.

When I visited Perimeter in 2022, Shayan insisted on a selfie with both his PhD advisors.

This June, an illness took Ray earlier than expected. We physicists lost an intellectual explorer, a co-founder of the quantum-computing community, and a scientist of my favorite type: a wonderful physicist who was a wonderful human being. Days after he passed, I was holed up in a New York hotel room, wincing over a web search. I was checking whether a quantum system satisfies certain tenets of quantum error correction, and we call those tenets the Knill–Laflamme conditions. Our community will keep checking the Knill–Laflamme conditions, keep studying quantum gates implementable with linear optics, and more. Part of Ray won’t leave us anytime soon—the way he wouldn’t leave a nobody of a master’s student who needed a conversation.

1For the record, some of the most rigorous researchers I know work in Cambridge’s Department of Applied Mathematics and Theoretical Physics today. I’ve even blogged about some

2As I still am, thank you very much.

July 25, 2025

Clifford JohnsonFantastic Collaboration!

Well, I can now officially mention that I've been part of the filmmaking team (in a way) working hard to bring you an enjoyable and interesting Fantastic Four movie! I think it has been about two and a half years (?) since this all began. This was a nearly perfect model of how science consulting can work in film. I worked with everyone, wherever I was needed, with the director, writers, producers, director of photography, VFX teams, set design, and so on. They made me feel welcome and part of whatever creative team I was talking to, which was great. They were open to lots of ideas right from when they were starting out thinking about tone, story ideas, and so forth, right through to final (key) tweaks right at the end of the process as recently as mere weeks ago.

It began early on with with having great conversations Matt Shakman and his writing team about the fact that Reed Richards is first and foremost a curiosity-driven physicist (and so quite different from the engineer we have in Tony Stark that we see RdJ bring out so well), and how things like his dedication to his work (and his outlook on things that comes from such work) might play out in terms of family dynamic, personal relationships, etc., - Without it turning into the tedious cliches about scientists somehow not being able to navigate the world of human relationships. Obviously, I could speak to this as a physicist who works on precisely the things Reed works on, as well as a family man, and as well as someone who remembers that it's still all about telling a story. And there are so many stories to tell at that intersection... Anyway, I think these early conversations (as well as suggestions I made in many sets of notes along the way) helped inform (even if only a little bit? who knows?) what Pedro Pascal brought to the character. This aspect of the film is one of the things I'm most pleased about seeing up on screen.

Beyond that, you'll see lots of things I gave them that I'm also delighted to see made it to the film, in many scenes. This includes (but not limited to!): [...] Click to continue reading this post

The post Fantastic Collaboration! appeared first on Asymptotia.

David Hogghow significant is your anomaly?

So imagine that you have a unique data set Y, and in that data set Y you measure a bunch of parameters θ by a bunch of different methods. Then you find, in your favorite analysis, your estimate of one particular parameter is way out of line: All of physics must be wrong! How do you figure out the significance of your result?

If you only ever have data Y, you can't answer this question very satisfactorily: You searched Y for an anomaly, and now you want to test the significance. That's why so many a posteriori anomaly results end up going away: That search probably tested way more hypotheses than you think it did, so any significances should be reduced accordingly.

The best approach is to use only part of your data (somehow) to search, and then use a found anomaly to propose a hypothesis test, and then test that test in the held-out or new data. But that often isn't possible, or it is already too late. But if you can do this, then there is usually a likelihood ratio that is decisive about the significance of the anomaly!

I discussed all these issues today with Kate Storey-Fisher (Stanford) and Abby Williams (Chicago) today, as we are trying to finish a paper on the anomalous amplitude of the kinematic dipole in quasar samples.

July 24, 2025

David Hoggfinding emission lines (and other oddities) in hot stars

I showed my robust spectral decomposition (dimensionality reduction) and residuals to the MPIA Binaries group today. There was much useful feedback (including that my H-gamma was actually H-delta; embarassing!). One comment was that the model isn't truly a causal separation between star and lines, so there will be some mean lines in the star model; lines aren't entirely outliers. That's true! The group suggested that I iterate to remove stars with lines from the training set.

After the meeting, I implemented some of that, but problems like this have a pathology: If you carefully remove stars with high residuals at some wavelength, then the training data will be deficient, or low, at that wavelength. And then the model will go lower, and then more stars will have excess at that wavelength and: Disaster. So when I implemented, I required a 2-sigma deviation, and I removed both high and low outliers. I don't know if this will work, but I am testing now.

July 23, 2025

Mark GoodsellEntangled colliders

There are several interesting papers on the arXiv today. One of them, arXiv:2507.15949, involves my former PhD supervisor. It's on the subject of Quantum Entanglement at collider experiments, and relates back to a paper of his from 1992 that I didn't know about (there's a great line in the new paper where the authors complain that their earlier paper was ignored). (Quantum) Entanglement is the phenomenon where two or more particles are in a special state so that their properties are related, but we don't know what those properties are until we measure them. In Quantum Mechanics we would say that the actual state is not decided until we measure them, and this leads to 'spooky action at a distance' because by measuring one particle we appear to set the corresponding property of the other. An alternative explanation would be that there is some hidden quantity or 'hidden variable' where both particles secretly know all along what state they are in. However, surprisingly it's possible to discriminate between these two cases, and set up quantitative tests known as 'Bell inequalities': you can make a measurement and, if the result of that measurement is less than a certain value, then a hidden variable theory cannot explain it. Experiments to test this using photons at low energies were performed in the early 80s by Alain Aspect and others that violated Bell inequalities and thus confirming the Quantum Mechanical interpretation. 

In recent years, experimentalists have become interested in performing similar tests using different particles at higher energies; it is legitimate to ask "is this true for fermions?" or "does this break down at high energy?" Apparently similar questions were asked in the early 90s at LEP where electrons and positrons were collided (instead of protons at the LHC) and the 1992 paper pointed out that they were not really testing Bell Inequalities. The new paper revisits the older argument, and applies it to the new case of e.g. proton collisions producing a top-antitop pair. They argue that the quantity of interest for the Bell Inequality is the spin density matrix:

But what can actually be measured is the differential cross-section (the rate of production of particles in a certain angular volume):

The symbols B and C appear in both expressions: when performing experimental tests of Bell inequalities they are identified with each other. Since the differential cross-section can be measured, the measurement for the Bell Inequality can then be made and tested. However, the authors of the new paper claim that, in order to identify the two sets of symbols, it is necessary to use Quantum Field Theory: the second equation is a prediction based on QFT from the first. In other words, the logic is circular, and Quantum Mechanics has been assumed -- so it's not surprising that the Bell inequality is violated!

I haven't worked on this topic myself, so it will be interesting to see if there is some pushback from the authors of papers such as arXiv:2003.02280 (who proposed such top-antitop studies). 


Fermi decay constant -- at three loops!

 I also want to point out arXiv:2507.15946 by Stephen Martin, who has performed a three-loop computation of the decay rate of the muon in the Standard Model at three loops. This quantity is incredibly important; it is measured very precisely, and so we use it to extract the underlying parameters of the Standard Model -- or, any theory beyond it. But since it's a complicated process, this is a tricky computation, even at low loop order. The results in this paper will be useful for all sorts of calculations, such as extracting the Higgs boson's self-coupling -- and working out whether the universe is metastable!

July 22, 2025

David Hoggwrote like the wind; frequentist vs Bayes on sparsity

My goal this year in Heidelberg is to move forward all writing projects. I didn't really want to start new projects, but of course I can't help myself, hence the previous post. But today I crushed the writing: I wrote four pages in the book that Rix (MPIA) wants me to write, and I got more than halfway done with a Templeton Foundation pre-proposal that I'm thinking about, and I partially wrote up the method of the robust dimensionality reduction that I was working on over the weekend. So it was a good day.

That said, I don't think that the iteratively reweighted least squares implementation that I am using in my dimensionality reduction has a good probabilistic interpretation. That is, it can't be described in terms of a likelihood function. This is related to the fact that frequentist methods that enforce sparsity (like L1 regularization) don't look anything like Bayesian methods that encourage sparsity (like massed priors). I don't know how to present these issues in any paper I try to write.

Justin WilsonWelcome to the Quantum World: Where Certainty Ends and Possibility Begins

1. The Classical vs. Quantum World

In our everyday experience of the world, things have precise positions, speeds, and outcomes. You throw a baseball—you know where it’s going. But when we zoom in to the world of atoms and particles, things get weird — and the rules change.

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support my work.

2. The Probabilistic Nature (Uncertainty and Superposition)

🗨️ Metaphor:

"Imagine flipping a coin, while it is spinning in mid-air, it spins in mid-air being both at heads and tails at the same time, with the probability of being heads or tails is still 50-50. At this point, if we want to describe the state of this system (the coin), it would be a combination of both heads and tails — until you look, and then you can say whether the coin landed on heads or tails. That’s how particles behave in the quantum world: they exist in a state made of both heads and tails, a superposition of states, until they’re measured.

🎯 Main idea:

  • Quantum Particles don’t have exact positions or velocities—just probabilities.

  • Measurement collapses the particle’s wavefunction to a definite value.

Let’s look more closely at the idea that Particles behave probabilistically

In classical mechanics, we think of a particle as a tiny object with a definite position and velocity at any time. But in quantum mechanics, particles like electrons that are described by a wavefunction, a mathematical function that tells you the probability of finding the particle in different places. You can think of the particle not as a dot but as a fuzzy cloud, where he denser the cloud in one spot, the more likely the particle is to be found there.

This is why we say: "Particles don't have exact positions or velocities—just probabilities."

🎵 The Wave Nature of Matter

In our everyday life, we see systems that exhibit wave properties. Things like sound waves, water waves (surface waves), waves on a cable (vibrating), or if you live in certain places, you may experience seismic waves. These are all classical physics examples that are described by wave equations, where the disturbance propagates through a medium or field, transferring energy without necessarily transferring matter.

For example, when waves meet (i.e., waves in water), they combine through a process called interference. This can take a few forms:

· Constructive Interference: When the crests (high points) and troughs (low points) of two waves line up, they reinforce each other, creating a larger wave. Think of two ripples on a pond colliding and forming a bigger splash.

· Destructive Interference: When a crest meets a trough, they cancel out to some extent—sometimes completely—resulting in a smaller or flat wave.

This blending of energy is happening constantly in light, sound, water waves, and even quantum systems.

Below in Figure 1, is an example of superpositions of waves. The top image highlights full constructive interference and the bottom image shows destructive interference. You can see that the maximum of the two waves is 1 and its minimum is -1, where 1 and -1 are called the wave's amplitude. For these two points, for complete constructive interference, the superposition of these waves yields 2 (superposition means at each position point you add the two waves together) for the maximum and -2 for the minimum. For complete destructive interference, you can see the waves when at each point you add them together (superposition), completely cancel out (equal 0). This situation is often called completely out-of-phase . Using the same two points as in our constructive interference example, you now see that wave 1 equals 1 and wave 2 equals -1. In fact, for all the points, the two waves are equal but of opposite sign (meaning one is positive, say +1, and the other is -1). The superposition of these two waves produces 0 for all points.

Figure 1: Top image showing complete constructive interference, while the bottom image displays complete destructive interference.

Below in Figure 2, the waves are slightly shifted along the position axis (x-axis). Using our same points as before, you can see that the superposition wave doesn’t quite equal 2 and -2; they are less than 2 and greater than -2 (-2 is less than -1.9, say, meaning -2 does not get more negative, it is heading upwards towards 0). This is because each wave’s maximum and minimum values occur at different points in space, and this is true for the values of the superposition wave at all points in space. Imagine you fix wave 2, and you slowly pull wave 1 to the right (wave 1 could be referred to as phase-shifted relative to wave 2). The superposition wave continues to have positive values and negative values going towards 0. Once the maximums of wave 1 line up with the minimums of wave 2, the superposition wave is 0 for all points. This is the complete destructive interference as we saw in Figure 1. Now, if you continue to pull wave 1 to the right, the superposition wave starts growing, and if you keep pulling to the right, it will reach the complete constructive interference pattern like in Figure 1.

Figure 2: Two waves shifted relative to one another along the x-axis (position axis).

Notice the superposition wave (like the other waves) starts to repeat the pattern. The point where the pattern repeats itself would define the superposition wave’s wavelength 𝛌. Now imagine, if you had lots of waves where some are shifted relative to our wave 1, at some points in position, we will get a maximum amplitude resulting from constructive interference but necessarily complete constructive interference, giving the highest point of a wave (crest of water wave), while for others, we may get destructive interference, leading to the minimum amplitude (trough of a wave) and other intermediate amplitude that help to make-up the entire wave. Hopefully, this simplistic model helps us to understand how waves form and how you can get a big wave from many small waves.

Another feature of waves is that they have a wavelength that describes how far they propagate in space before repeating the same pattern over and over. If you remember what the mathematical functions, sine and cosine, they are waves that repeat in space and have a wavelength. Now the important part is that the momentum, p, of these waves is inversely proportional to their wavelength, that is, p=1/𝛌. So if you have a short wavelength, you have a large momentum, and vice versa.

These waves follow classical equations — disturbances that move through a medium, transferring energy. But in quantum mechanics, the wave isn't a ripple in water or air — it’s a probability wave.

Now comes the key idea: wave-particle duality. Particles act like waves. And waves behave very differently from particles in one crucial way:

A wave that's localized in space (i.e., sharply peaked in position) must be made by combining many different wavelengths. Think of a big wave in the ocean; it is formed by lots of waves coming together to form this big wave. This combining of waves also means you have a wide range of momenta.

Correspondingly, a wave with a defined momentum (i.e., well-defined momentum) must be spread out in space.

For example, let’s look at music and a pure note on a tuning fork (single frequency = defined momentum) lasts long but is hard to pin down in time (spread out). However, a short drumbeat is localized in time (defined position) but contains a spread of frequencies (momentum uncertainty).

For example, let’s look at music and a pure note on a tuning fork (single frequency = defined momentum) lasts long but is hard to pin down in time (spread out). However, a short drumbeat is localized in time (defined position) but contains a spread of frequencies (momentum uncertainty).

This is a fundamental mathematical property of waves called the Fourier transform. A Fourier transform contains both sine and cosine, just as waves, but is a more complicated function that involves complex numbers. The point about the Fourier transform is that you can obtain sine and cosine from it.

3. The Heisenberg Uncertainty Principle: Knowing Less to Understand More

One of the most famous — and misunderstood — ideas in quantum mechanics is the Heisenberg Uncertainty Principle.

It’s often summed up like this: You can’t know both where something is and how fast it’s moving — at the same time — with perfect precision.

At first glance, that sounds like a problem with our measuring tools, as if we just need better microscopes or sensors. But that’s not it.

This principle isn’t about technological limitations — it’s a fundamental property of nature.

What does it mean?

In classical physics, if you know where a car is and how fast it’s going, you can predict exactly where it’ll be a few seconds later. But in the quantum world, if you try to pin down the position of a particle more precisely, you automatically become less certain about its momentum (its speed and direction) — and vice versa.

It’s not because the particle is misbehaving — it’s because particles aren’t like tiny billiard balls. They behave like waves, and waves don’t have sharp edges.


🌀 Wave Metaphor

Think of a musical note. If a sound wave is spread out in time — like a long, steady tone — it has a very precise frequency (pitch). But if it’s a short, sharp “ping,” its frequency becomes less certain. You trade time for pitch.

In the same way, if a particle’s wave is sharply localized in space (you know where it is), the range of its momentum values must broaden. If the wave is spread out (you don’t know exactly where it is), the momentum is better defined.


🔬 So what’s uncertain?

It’s not that the particle is jittering around randomly. Instead:

  • Before measurement, a particle’s position and momentum are both described by a range of probabilities.

  • The more tightly you narrow one, the more uncertain the other becomes.

The Heisenberg Uncertainty Principle can be written down as,

𝚫p𝚫x ≤ ℏ/2

  • 𝚫x is the uncertainty in position

  • 𝚫p is the uncertainty in momentum

  • ℏ is Planck’s constant (a very small number)

Let’s try to understand this formula a little better. In quantum mechanics, particles like electrons aren’t just little dots — they also act like waves. This means we describe them with wave packets, which are like short-lived ripples or pulses spread out over space.

To make a wave packet that’s narrow in space (so we know roughly where the particle is), we have to combine many different waves (i.e., sine waves) with various wavelengths and frequencies (think back to our above example of waves).

That’s because a single sine wave, for example, stretches out infinitely — it doesn’t give you a clear position. Only by mixing waves with different wavelengths (and therefore different momenta) can we build a localized bump.

So: Precise position → requires many different wavelengths → high momentum uncertainty.

Now reverse it. If we only use one sine wave, it has a very clear wavelength (momentum), but it stretches out forever — the particle could be anywhere.

So: Precise momentum → means the particle is spread out → high position uncertainty.

This trade-off is at the heart of the uncertainty principle:

𝚫p𝚫x ≤ ℏ/2

Here, 𝚫x is the uncertainty in position, 𝚫p is the uncertainty in momentum, and ℏ is a very tiny constant from quantum physics.

The key message: > The more precisely you know where something is, the less precisely you can know how fast it's going — and vice versa.

Imagine building a short splash on a pond with water waves (see Figure-3):

  • A small, sharp splash uses many different ripple sizes (frequencies).

  • A pure, smooth ripple has just one frequency but spreads out.

That’s the uncertainty principle in action, hiding in the rhythm of waves.

Figure 3: The left figure shows the sharp splash, while the right figure illustrates the smooth ripple.

So what that tells us is that as we become more and more certain about the location of a particle (𝚫x is getting smaller and smaller, heading to 0), 𝚫p is getting larger and larger, heading to ∞. This tells us that if we knew x exactly, then we would not know the momentum p of the particle, since the uncertainty 𝚫p is infinite.

The Core Idea:

You can’t precisely know both where something is (position) and how fast it’s going or in what direction (momentum) at the same time. The more accurately you try to measure one, the fuzzier the other becomes.

🧠 Everyday Analogy:

Imagine you're trying to photograph a speeding car at night.

  • If you use a fast shutter, you can see exactly where the car is, but the picture will be blurry — you can’t tell how fast it was going.

  • If you use a slow shutter, you get a motion blur — which tells you how fast it was moving, but now you don’t know exactly where it was.

That’s the uncertainty principle in action: precision in one area means fuzziness in the other.

Again, this isn’t just a limitation of our instruments — it's a fundamental property of nature. It's like the universe itself has this built-in fuzziness at tiny scales.

This principle also tells us why electrons just don't spiral into the nucleus of an atom.

Because you can’t precisely know both the position and momentum of a particle at the same time.

If an electron got too close to the nucleus, its position would be very well known (i.e., tightly confined in space). According to the uncertainty principle, this would mean its momentum becomes highly uncertain. Because the kinetic energy is directly calculated from the momentum, and since you have large momentum fluctuations, you will have large kinetic energy.

This tells us that confining the electron too tightly costs energy — a lot of energy. That energy cost balances out the attractive pull of the nucleus. The result? The electron occupies a fuzzy “cloud” of most likely locations (remember it is based on probabilities)— what we call an orbital — and it doesn't just fall in.

This quantum balancing act gives rise to stable atoms, the periodic table, chemistry, etc.

Wave-particle duality

Wave-particle duality is one of the most astonishing ideas in modern physics. It says that tiny things—like electrons and light—can behave like particles and waves, depending on how you look at them.

  • Waves (like ocean waves, or ripples in a pond, or even sound waves) are spread out, continuous disturbances. They travel, they can interfere with each other (creating bigger or smaller waves), and they bend around corners. You can't point to "one wave" and say it's at a single, precise location.

  • Particles (like a baseball, or a tiny pebble) are distinct, localized objects. They have a definite position, mass, and can be tracked as they move from one point to another.

The Classical Difference: In our ordinary experience, something is clearly either a wave or a particle. Never both.

🌍 In the Classical World

In everyday experience:

  • Objects are either particles (like baseballs) or waves (like sound or water ripples).

  • Particles have defined positions and travel along clear paths.

  • Waves are spread out, overlap, and interfere, but they don't "exist" in a single spot.

Think of throwing a rock into a pond—either you're dealing with the rock or the ripples it creates, never both at once.

⚛️ In the Quantum World

The Quantum Twist: Wave-Particle Duality

But when we zoom down to the incredibly tiny, fundamental level of reality – the quantum realm – things get weird. Particles like electrons, and even light itself (which we classically considered a wave), don't always fit neatly into one category. This is wave-particle duality:

  • Light, for instance, can behave like a spread-out wave (which is why it can create interference patterns, just like water waves). But it can also act like a stream of tiny, discrete particles called photons (which is how it knocks electrons off a metal surface in the photoelectric effect, acting like tiny billiard balls).

  • Similarly, electrons (which we think of as particles making up atoms) can, under certain experimental conditions, exhibit wave-like behavior, creating interference patterns as if they were spread out and passing through multiple places at once. Yet, when we try to pinpoint their location, they act like a localized particle.

This means a single electron, shot toward a double slit, doesn't just go through one slit—it behaves as if it explores all possibilities at once, producing an interference pattern typical of waves.

🤔 So What Does This Mean?

The amazing part is that a quantum entity isn't just sometimes a wave and sometimes a particle. Instead, it possesses both wave-like and particle-like properties simultaneously, and the act of observation or the type of experiment we perform determines which aspects we will observe. You can't observe both characteristics at the same exact time in the same experiment.

This seemingly paradoxical idea is a cornerstone of quantum mechanics and is absolutely essential for understanding how the universe works at its most fundamental level. It underpins all modern technologies from lasers and transistors to medical imaging and the very concept of quantum computing.

The objects aren't just "here or there"—they are probabilistic ripples, until observed.

Wave-particle duality is nature’s way of whispering: “The world is more nuanced than it seems.”

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support my work.

July 18, 2025

Justin WilsonScientists Discover a New Phase of Game Show!

You’ve seen the headlines: “Scientists Discover a New Phase of Matter!” They usually go something like, “You’ve heard of solids, liquids, and gases—but now there’s a fourth (fifth) phase: [insert buzzword here].”1 You might think about these phases in terms of temperature: heat up ice and it melts, a phase change! And yes, that is a phase transition. But temperature is just one knob we can turn, and phases are far richer than just “solid, liquid, and gas.” In fact, new phases are surprisingly common, and to understand why, let’s play a little game.

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support our work.

What is the Percolating Phase?

Imagine you’re on a game show called Are You Likeable? (the least-likeable game show). The rules of the game are simple:

  1. You stand on the stage and try to win over the audience.

  2. Each audience member votes whether they like you or not.

  3. But the twist: votes aren’t tallied—they control a system of pipes above your head.

That system of pipes looks something like this2

A game where you get soaked or remain dry based on whether the audience votes, which randomly turns a spigot on or off. Image generated by ChatGPT.

Each “like” turns a spigot off, stopping water from flowing through one pipe in a grid overhead3. Once voting ends, water is dumped into the system. If it can find a path to the bottom, you get soaked. The better your “likeability,” the less likely spigots open a path for water to flow and the drier you stay. That’s your prize for this game show (and hey, you also get the knowledge that people out there like you).

This system models a type of phase transition known as percolation4.

But where is the phase transition?

Aside from asking when my game show will be green-lit5, we can ask: When are you most likely to get wet? If the audience is huge, and fewer than 50% of them like you, it’s nearly guaranteed that the water will find a path—you’ll be soaked. But if more than 50% like you, chances are good that you’ll stay dry.

This is a phase transition6: a sudden change in the system’s behavior. For a moderately sized audience, it looks something like this:

The phase diagram for when your likeability gets you soaked in this game show. The line represents your chance of getting soaked.

Your probability of getting soaked forms a curve that sharpens up with increasing audience size—becoming a near step at 50%. That is known as the percolation threshold.

This is hard to visualize, though; luckily this problem admits some very nice pictures. For instance, here is the problem with a large number of pipes:

Animated gif of the water flowing through with certain “Likeability” scores. (Refresh if the animation stopped.)

If a pipe is blue, water is in it, and all the blue clusters flow down from the top. Notice what happens around 50%: even though spigots are randomly being turned off, the flow from top to bottom is entirely stopped around this value. Something is happening within the system that is allowing water to pass through on one side of the transition and not the other.

A closer look at the transition

To dig deeper, we simulate what happens at this threshold. Each spigot is either open or closed (randomly determined). If we visualize the grid (say 1024×1024 spigots), it looks like visual static: black and white dots with no obvious pattern7:

But now let’s color each connected cluster of open spigots—where water could flow and fill up this section of pipes. Suddenly, structure emerges. Some clusters are small, others large. If one spans from top to bottom, water flows and we’re in the percolating phase. If not, we’re in the non-percolating phase. At the transition (within the static above, we get this for the twelve largest clusters:

At exactly the percolation threshold (50% for the pipes above), there’s no single dominant cluster, but also no clear typical size. Instead, there’s a wide distribution of cluster sizes. The critical state behaves differently than either phase.

Scale Invariance: A Hallmark of Criticality

Let’s zoom out. Suppose we double the grid to 2048×2048.

The largest clusters are definitely larger—the largest here is 400,000 pipes/spigots large, while for the previous 1024×1024 case the largest was 180,000 large—but the pattern still looks… the same. We doubled the size, but if we slightly blur our vision, we cannot distinguish these two plots (even though one is quadruple the area of the other). Look at 512×512—even that looks similar:

You would be hard-pressed to say which one is larger if you blurred your eyes. This problem is apparent even down to 256×256 or 128×128.

This is called scale invariance—there is no characteristic length scale at the phase transition. It’s one of the defining features of what are known as second-order phase transitions.

It also explains why, at the threshold, you have a 50% chance of getting soaked. The largest cluster might span the system, but it might just as well fall short. There’s no guarantee either way.

Fractals in the Flow

These clusters within the above pictures don’t look like regular 2D structures or even 1D lines; they are, in fact, fractals. They aren’t exactly self-similar but they do behave the same at different scales. They fill space with a fractal dimension: not quite 1D, not quite 2D. In two-dimensional percolation, the clusters have a dimension of 91/48 ≈ 1.896—a universal number shared by all systems in this class, regardless of lattice type or other microscopic details.

This is part of the beauty of percolation: It shows us visually the underlying mathematical structure and even reveals some universality of phase transitions.

Why This Matters

Percolation is just one example, but it captures the essence of what physicists mean when they talk about “phases of matter.” It isn’t always about exotic particles or extreme temperatures turning gas into plasma. Sometimes, it’s about whether a liquid can find its way through a series of pipes. It’s about symmetry, structure, and emergence.

You’ve experienced water’s phases: ice, liquid, steam. But nature offers many more—some with no neat label like “solid” or “gas.” The theory of phase transitions explains them. Percolation is a window into that wider world.

1

What’s funny to me about this is how we use ice/water/vapor to expound on this. But water’s phase diagram is complex and deserves its own post. One point of interest: water and water vapor can be smoothly connected to each other without going through any phase transition. That and there’s something like 20 phases of ice.

2

Forgive me the ChatGPT weirdness: the drip on the left pipe and the weird long pipe on the right are just LLM quirks. Kind of like how hard it is to get AI to draw a full wine glass. (Or maybe now it can?)

3

Audience members can’t influence each other here. Assume the spigots are randomized and a stern librarian keeps everyone silent.

4

Technically, this is bond percolation on a square lattice.

5

NBC, call me!

6

A second-order phase transition is one where the change is continuous, but its derivatives (like heat capacity or cluster size) diverge. Percolation is a particularly visually clean example.

7

I have switched from bond percolation to site percolation to make plotting and cluster finding easier. The universal features do not depend on this.

July 15, 2025

Mark GoodsellThe Ant Mill

Jesper Grimstrup kindly sent me an electronic copy of his new book, The Ant Mill. He was also kind enough to give me some feedback on a first version of this review.


It has a foreword by Peter Woit, who has commented briefly about the book on his blog; the author also has a substack. The subtitle is 'How theoretical high-energy physics descended into groupthink, tribalism and mass production of research' so you would expect it to be the sort of thing that I would take a strong objection to. However, I am the sort of person who likes to read things that challenge me; the only thing that gets under my skin in this book is attacking the whole of academia in public.

The story is an interweaving of the author's personal experiences in academia with his general observations. This personal story and his experiences are interesting, much like I expect those of everyone who has spent several years in academia would be. And he has clearly spent a lot of time thinking about thinking about research. I love meta-activities of this sort; the best example that I know of is You and Your Research by Hamming, which I stumbled on as a postdoc. Indeed, the existence of these sorts of things that are shared by young researchers is actually evidence against the central thesis of Grimstrup's book.

The market attacking High-Energy Physics seems to be burgeoning. On the one hand Hossenfelder believes that we have become 'lost in math,' and on the other Woit believes we are not mathematical enough; both attack string theory as a failed program. Grimstrup's book is in the mathematical camp, with the novelty that he piles scorn on all popular approaches to quantum gravity, in particular loop quantum gravity and noncommutative geometry, since he has come into closest contact with them. His observations about string theorists are mainly about the shoddy way that he was treated during his time at the NBI, with several egregious examples of bad behaviour. We are lead to conclude that it is not just string theorists who have formed a closed tribe, but that there are several such groups crowding out innovation.

Grimstrup refers to his own research program and gives examples of how it has just generally been ignored within academia. For example, he starts the book with a copy of a grant application by a 31-year-old Niels Bohr for an entire institute, and contrasts this with a grant application of his that was refused that effectively ended his career within academia (my understanding is that at the NBI in Copenhagen it is common to ask for and obtain grants to pay your own salary and prolong temporary contracts). He writes that he does not do this to compare himself to Niels Bohr, but inadvertently this is the impression I got from the book -- that he was doing this not in a self-aggrandising way, but in the sense that you can almost feel his frustration coming through the pages that his expectations did not meet reality. It seems like bait at times, inviting anyone who disagrees with the general thesis to attack him personally. Instead, I will have a look at his papers with an open mind, after writing this review, and keep my thoughts on them to myself.

The book made me think of how many of us enter academia. We grow up reading popular science accounts idolising physicists from a century ago. And it made me think more of the self-actualisation messages that were rammed down all our throats in the popular culture in the 80s and 90s: follow your dreams, stick to your principles, be true to yourself, this is the most important thing in life and you shouldn't worry about money, just be happy. And: working hard and getting good grades is the way to get to the top. The problem is that this is largely obsolete: it's based on the world that existed post world war two when there was a scarcity of labour and an economic boom. Then -- if you were from the right background and your face fit -- you could work hard, get a PhD and walk into a permanent academic job (yes this is a caricature). Science was respected and so were scientists; high-energy physics was at the top of the tree because of the connection with technological advancements and nuclear weapons. That world doesn't exist any more; while in many ways for the better, it is undeniable that we live in a world of much greater competition and public skepticism about science is increasing.

The scientific community has expanded, as has the population; and more importantly education throughout the world and global travel and communication has meant that the number of people around the world who are involved in research is much greater than it was. Grimstrup notes that increasing the size of the academic community has led to fundamental changes of behaviour: professionalisation of research and group think, and that this leads to an increasing incentive to work on mainstream topics. He has done bibliographical research to demonstrate this (in papers and presented in the book). It is clearly true that the Matthew effect exists in many branches of society, and therefore also in academia; governments wanting to exert some form of oversight in exchange for the funds that they provide has definitely led to changes in incentives for researchers. One aspect of this is that it is hard to judge the work of people from other fields, but we are required to do so; and then it is difficult to argue with quantitative measures such as number of papers, citations, h-indices. Then of course the measure becomes the target for certain people. 

Grimstrup rails against all these changes; he clearly believed that the correct thing to do for an aspiring researcher would be to work on their own ideas, stick to their principles and not compromise. They should work for a long time, in isolation, on a major paper, put it on arxiv.org and the next day their colleagues would read it and ask interesting questions about them. Fame and fortune would follow. The thing that shocked Grimstrup was that not only did people not even care about any papers he posted, a young competitor even once told him some ideas are simply not worth pursuing even though they may be interesting. For sure, this is horrible and shocking behaviour, and does not reflect well on the anonymous person who said it.

For my part I am still naive enough to think that if new ideas are good, someone will recognise them as such, and network effects will make them known. I know that many researchers already think more deeply about what they are doing than he gives us credit for: and we discuss it, during seminars, over a drink with colleagues, in the coffee-breaks of conferences, during our annual or five-year reviews, or in grant applications. When I discussed this review with a string-theorist colleague they remarked "of course we know the situation sucks!''  I think Grimstrup is therefore wrong to tar everyone with the same brush: the diversity in our community has increased greatly with time, and this means that there are indeed strong incentives to take a risk on a novel idea, because the rewards of opening a new research direction are immense! Being the originator of an idea, or the first to recognise the merit in even an old forgotten idea, can yield tremendous results and even greater recognition nowadays thanks to the same effects. Hence, starting a new field, or even a subfield, is something that most researchers aspire to; the rewards for doing so are even greater now than in times gone by, and the evidence that this is possible is even given in this book: the existence of several communities working on different approaches to quantum gravity. He argues that these are now old and stale, but my point is that the way that they were able to take root at all is an example of how this can happen. There are many subfields that have sprung up more recently, and in other branches of HEP there are of course many examples. Nowadays things can change very quickly: a new good idea will be very rapidly jumped on once it is recognised, and people are constantly on the lookout. 

Grimstrup also, like Lee Smolin, divides researchers into visionaries and technicians. He then complains that the technicians have taken over, with lots of disparaging comments about them digging endless holes. He then complains that there is an incentive to collaborate in modern research, only collaborators survive in the system: he has evidence that being a lone wolf is a poor survival strategy. He believes that we should work on our own; yet at the same time visionaries need to collaborate with technicians. I found this very jarring. Other than the facile placing of people into boxes, he is overlooking the benefits of collaboration -- his opinion is that it is just about inflating the number of papers one person can sign (and for sure there are people who cynically do this). But to me, discussing with other people, even just explaining something, is often the quickest way to generate genuinely new ideas or solutions to problems that we may never have come up with alone. At the same time, there are plenty of people who do write papers alone; to take a leaf from his book and share a personal story, I once had a comment on a postdoc application that I had no single-author papers and therefore did not demonstrate independence. Hence, there are incentives and a good reason for young researches to work alone sometimes. I then wrote a single-author paper, as I have occasionally done since (and got the fellowship next time I applied); I would agree that there is a pleasure and some advantages in doing this, but to do this all the time would mean I would risk missing out on lots of new ideas and other perspectives, as well as the pleasure of regular interactions with collaborators, and it would also limit the scope of my projects, where I benefit from others' expertise. Or collaborations may just be working with a student, pursuing my ideas (hopefully they contribute some of their own!) and imparting my knowledge in the process. This is why I do not think that encouraging people to predominantly cloister themselves away to work alone for a long time is the most productive or healthy one. 

The book also has a very narrow focus as to the goal of high-energy physics. For the author, the quest is a "the next theory," but in essence this means a theory of quantum gravity, which he acknowledges would be far from being able to be tested with any present or near-future data. Otherwise, we should look for a mathematically rigorous definition of quantum field theory; he hopes these will be one and the same thing. This latter problem has proven to be both very hard and not obviously useful -- it is certainly not obvious that the solution should even be unique, for example a theory of strings would cure ultra-violet divergences, and the question of whether strings should be necessary for such a theory is one that I know people have tried to explore. I also recently attended a talk by Michael Douglas where he reviewed recent attempts on rigorous QFT, so it is a subject that is regarded as important but very difficult, and still being explored by a small number of people. Regarding quantum gravity, some people in the community have taken the opinion that if you have no data, it is not a good problem, and are working on other things. Or people try to make contact with data using e.g. EFT approaches to measuring quantum effects of gravity. The string theory community might say that we do have a theory of quantum gravity, in fact we have a landscape of them, and try e.g. to use it to answer questions about black hole information. But at the same time some people then complain that the leading string theorists have moved on to other things: there are lots of important open fundamental problems, and we just do not know how they are interlinked, if at all!

Grimstrup's insistence that the solution to what he sees as problems is to shrink competition and also encourage research outside of academia, reminded me of another Dane, subject of another book I read recently: king Cnut, famous for (presumably apocryphally) standing on the beach in front of his ministers and commanding the tide to turn back. Otherwise Grimstrup hopes for a crisis, perhaps one provoked by his book. He explicitly states that he does not want to fuel the anti-establishment or ant-academic movements, but I suspect that the only crises we might suffer would not be good for the field.  Perhaps one is already taking place in the US; perhaps people will take his message to heart despite his protests and start a DOGE-style decimation of research. Necessarily, in science we mark our own homework: only other scientists are capable of judging the claims of their peers. If we start opening this up to question then we will only end with government appointees deciding what are acceptable topics and directions, or shutting public funding down altogether. What would be left over would surely be even greater competition for scarce resources.

For me, the solution to the problems in the book, to the extent that I agree with them, is to regularly remind ourselves that we should always maintain a childlike curiosity and not close our minds to new ideas and new possibilities. This is the message from the text of Hamming, and very well put in the writings of Feynman (who Grimstrup bizarrely dismisses as a technician compared to Bohr). Otherwise of course in science it is necessary to have a community spirit, to realise that we are all trying to make progress in the best way we know how, and to help each other do so; and it is necessary to maintain healthy competition as a motivator. But both conflicting instincts -- to compete and to group into communities -- are vital parts of human nature and denying this has been the mistake of utopians throughout history. 

I am also sure that many of the complaints that Grimstrup assigns to high-energy physics could also be applied to society more generally. So instead of trying to hold back or reverse the societal changes of the last century we should try to work with them as best we can. We have to accept that we live now in an attention economy; and this gives new opportunities: blogging, social media, writing articles in science magazines or popular press, etc. Since Grimstrup is now, interestingly, an independent scientist, perhaps tying his own research program so closely with his book is embracing the modern world at last, and creating a brand as a radical outside thinker, that will be attractive to private backers. He promotes the path that he has followed, crowdfunding his research or seeking support of patrons, as a possible path for the independently minded once they have completed their training in academia, and in this I wish him well: he is clearly serious, determined and sincere. But while this is now part of twenty-first century society, many people have noticed that this modern trend is a return to the nineteenth century (or even earlier, e.g. Leonardo da Vinci being invited to France by François 1) where a wealthy patron was the only source of funding. 



July 14, 2025

Clifford JohnsonThe Power of the String Equation

[More technical post follows.] I've been working on this project with (UCSB postdoc) Maciej Kolanowski on and off for a while now, but only in the last couple of weeks did I have the time to hunker down and help push the writing of the results to the finish. For your Sunday reading pleasure, it is already up on the arXiv here (it came out Thursday but I've been too busy to pause to post about it - partly because I've begun work on writing up the next paper in the backlog). The title is "Extended JT supergravity and random matrix models: The power of the string equation", and it is co-authored with Maciej Kolanowski.

In a way, it is a natural continuation of work I've described here from 2023 and 2024, described here and here. At a meeting at the Institute for Advanced Study in December 2023 I described in a talk (YouTube video here, look in particular from minute 35) something miraculous I'd discovered concerning capturing certain special supergravity (and black hole) behaviour using a random matrix model. The effective physics is [...] Click to continue reading this post

The post The Power of the String Equation appeared first on Asymptotia.

July 11, 2025

Justin WilsonTwo Dimensional Materials have gone crazy!

There are a ton of two-dimensional materials these days. You’ve probably heard of graphene, a single layer of carbon atoms arranged in a hexagonal grid.

a close up of a woven surface
In graphene, carbon atoms sit at the vertices of these hexagons. Photo by Andrew Draper on Unsplash

In 2018, everything changed when two layers of graphene were twisted to reveal superconductivity! The twist itself is interesting (I briefly discussed it in a previous post), but the key takeaway is that these materials now come with an extra knob for accessing new phases of matter. It’s remarkable. We can first think of these materials like Lego blocks:

blue, red, and white artwork
Photo by Omar Flores on Unsplash

Each layer is a different material: mix and match, and you might discover an exotic new phase. This “Lego” idea had already been in the air before 2018, but the physics since then has shown that it’s not just about stacking—we can twist too, creating not just patterns, but new ways for electrons to move.

Subscribe now

Two hexagonal layers twisted on top of each other, creating a moiré pattern.

We knew these patterns would occur, but we didn’t realize we could make it superconduct. Now we can stack and twist to great effect. Of course, twisted bilayer graphene isn’t about to revolutionize high-speed trains (it goes superconducting at only 4K1), but the way it goes superconducting is eerily reminiscent of higher-temperature superconductors. That means it might help us understand those other materials better.

And once people started twisting, they didn’t stop. We now have twisted multilayers of graphene, transition-metal dichalcogenide (TMD) bilayers2, and more. But it doesn’t end there; you can also apply magnetic fields, electric fields, and pattern the lattice in sophisticated ways. With all that in mind, here’s a short and incomplete survey of some of the exotic phases in these materials:

“Fractional… what now?”

All of these phases are exceptionally hard to understand and model. Some of the best minds in the field are actively working on them. One particularly exciting phase is the fractional Chern insulator, which could be useful for quantum computing.

But even setting aside applications, what’s astonishing is that all of these phenomena come from nothing more than electrons moving on a lattice and experiencing a few fields. Nature seems to treat electrons like Play-Doh, shaping them into wildly different quantum phases.

This is a deep and fundamental question: What can be accomplished using electrons alone?

1

That’s -452.47 degrees Fahrenheit.

2

To this day, I still can’t say the full name, so I just say “TMD.”