Planet Musings

November 28, 2025

Jordan Ellenberg“Not Vegan”

I saw a guy today wearing a hoodie that said “Not Vegan.” I’m not vegan either, not even close, but come on. This is like wearing a “Yankees Suck” shirt. I hate the Yankees too. But if what you put on your shirt is the thing you’re against, the thing you’re against owns you. Wear an hat for your favorite team and a shirt for your favorite meat. (Me: Orioles & ribeye steak)

November 27, 2025

Jordan EllenbergMathgiving

Happy Thanksgiving to all my American readers, and hey, everyone else too — gratitude is global.

In the US, this is the start of the season where people customarily think about their charitable gifts for the year, so here’s a few math-related places you could think of donating:

The Prison Mathematics Project supports mathematics study for inmates.

Bridge to Enter Advanced Mathematics (BEAM) runs math enrichment programs for low-income students.

And of course the good old American Mathematical Society continues to roll on supporting the American math community in innumerable ways, as it has for more than a century. You could donate money, or, if you’re a mathematician who’s not a member, you could join!

John BaezSummer Research at Topos

You can now apply for the 2026 Summer Research Associate program at the Topos Institute! This is a great opportunity.

Details and instructions on how to apply are in the official announcement.

A few important points:

• The application deadline is January 16, 2026.
• The position is paid and in-person in Berkeley, California.
• The Topos Institute cannot sponsor visas at this time.

These positions will last for 8 – 10 weeks, starting in June 2026 and ending in August. Each position will be mentored by Topos research staff or a select number of invited mentors. All positions are 40 hours/week, and the salary starts at $30-$50/hour.

There’s a research track and an engineering track. For the research track, possible topics include:

• Computational category theory using CatColab (Rust/Typescript skills recommended)
• Double category theory
• Categorical statistics
• Polynomial functors
• Interacting dynamical systems
• Hybrid dynamical systems, attractor theory and fast-slow dynamics
• Proof assistants, formal verification, or structure editors
• Philosophical and ethical aspects of applied category theory

For the engineering track, possible topics include:

• Delivery and support of mathematical technologies for various scientific disciplines and applications, and/or analysis, documentation, or guidance on their uses.
• Designing, implementing, testing, and maintaining software at the Topos Institute, in close collaboration with the research staff and in line with institute’s scientific strategy and mission.
• Contributing to developing the CatColab platform, including front end development in TypeScript and/or back end development in Rust. You might also contribute to the mathematical core, written in Rust, as your mathematical experience permits.

All positions require collaboration within a multi-disciplinary research environment. Each summer research associate will complete a specific Topos project, and will write a blog post by the last week of their employment. These projects may include an internal talk, software contribution, or paper. Go here to see the accomplishments of previous research associates.

Topos is committed to building a team with diverse perspectives and life experiences, so those with personal or professional backgrounds underrepresented at Topos are highly encouraged to apply. They are dedicated to shaping the future of technology to ensure a more equitable and just world, and believe that a technology that supports a healthy society can only be built by an organization that supports its team members.

John PreskillWhat distinguishes quantum from classical thermodynamics?

Should you require a model for an Oxford don in a play or novel, look no farther than Andrew Briggs. The emeritus professor of nanomaterials speaks with a southern-English accent as crisp as shortbread, exhibits manners to which etiquette influencer William Hanson could aspire, and can discourse about anything from Bantu to biblical Hebrew. I joined Andrew for lunch at St. Anne’s College, Oxford, this month.1 Over vegetable frittata, he asked me what unifying principle distinguishes quantum from classical thermodynamics.

With a thermodynamic colleague at the Oxford University Museum of Natural History

I’d approached quantum thermodynamics from nearly every angle I could think of. I’d marched through the thickets of derivations and plots; I’d journeyed from subfield to subfield; I’d gazed down upon the discipline as upon a landscape from a hot-air balloon. I’d even prepared a list of thermodynamic tasks enhanced by quantum phenomena: we can charge certain batteries at greater powers if we entangle them than if we don’t, entanglement can raise the amount of heat pumped out of a system by a refrigerator, etc. But Andrew’s question flummoxed me.

I bungled the answer. I toted out the aforementioned list, but it contained examples, not a unifying principle. The next day, I was sitting in an office borrowed from experimentalist Natalia Ares in New College, a Gothic confection founded during the late 1300s (as one should expect of a British college called “New”). Admiring the view of ancient stone walls, I realized how I should have responded the previous day.

View from a window near the office I borrowed in New College. If I could pack that office in a suitcase and carry it home, I would.

My answer begins with a blog post written in response to a quantum-thermodynamics question from a don at another venerable university: Yoram Alhassid. He asked, “What distinguishes quantum thermodynamics to quantum statistical mechanics?” You can read the full response here. Takeaways include thermodynamics’s operational flavor. When using an operational theory, we imagine agents who perform tasks, using given resources. For example, a thermodynamic agent may power a steamboat, given a hot gas and a cold gas. We calculate how effectively the agents can perform those tasks. For example, we compute heat engines’ efficiencies. If a thermodynamic agent can access quantum resources, I’ll call them “quantum thermodynamic.” If the agent can access only everyday resources, I’ll call them “classical thermodynamic.”

A quantum thermodynamic agent may access more resources than a classical thermodynamic agent can. The latter can leverage work (well-organized energy), free energy (the capacity to perform work), information, and more. A quantum agent may access not only those resources, but also entanglement (strong correlations between quantum particles), coherence (wavelike properties of quantum systems), squeezing (the ability to toy with quantum uncertainty as quantified by Heisenberg and others), and more. The quantum-thermodynamic agent may apply these resources as described in the list I rattled off at Andrew.

With Oxford experimentalist Natalia Ares in her lab

Yet quantum phenomena can impede a quantum agent in certain scenarios, despite assisting the agent in others. For example, coherence can reduce a quantum engine’s power. So can noncommutation. Everyday numbers commute under multiplication: 11 times 12 equals 12 times 11. Yet quantum physics features numbers that don’t commute so. This noncommutation underlies quantum uncertainty, quantum error correction, and much quantum thermodynamics blogged about ad nauseam on Quantum Frontiers. A quantum engine’s dynamics may involve noncommutation (technically, the Hamiltonian may contain terms that fail to commute with each other). This noncommutation—a fairly quantum phenomenon—can impede the engine similarly to friction. Furthermore, some quantum thermodynamic agents must fight decoherence, the leaking of quantum information from a quantum system into its environment. Decoherence needn’t worry any classical thermodynamic agent.

In short, quantum thermodynamic agents can benefit from more resources than classical thermodynamic agents can, but the quantum agents also face more threats. This principle might not encapsulate how all of quantum thermodynamics differs from its classical counterpart, but I think the principle summarizes much of the distinction. And at least I can posit such a principle. I didn’t have enough experience when I first authored a blog post about Oxford, in 2013. People say that Oxford never changes, but this quantum thermodynamic agent does.

In the University of Oxford Natural History Museum in 2013, 2017, and 2025. I’ve published nearly 150 Quantum Frontiers posts since taking the first photo!

1Oxford consists of colleges similarly to how neighborhoods form a suburb. Residents of multiple neighborhoods may work in the same dental office. Analogously, faculty from multiple colleges may work, and undergraduates from multiple colleges may major, in the same department.

Sean Carroll Thanksgiving

 (Apologies for the ugly blog format. We had a bit of a crash, and are working to get the template back in working order.)

This year we give thanks for a crucially important idea that can mean very different things to different people: information. (We’ve previously given thanks for the Standard Model LagrangianHubble’s Law, the Spin-Statistics Theoremconservation of momentumeffective field theorythe error bargauge symmetryLandauer’s Principle, the Fourier TransformRiemannian Geometrythe speed of lightthe Jarzynski equalitythe moons of Jupiterspaceblack hole entropyelectromagnetism, Arrow’s Impossibility Theorem, and quanta.)

“Information” is an idea that is everywhere in science and technology these days. From one angle it looks like such an obvious idea that it’s a bit startling to realize that information theory didn’t really come along until the work of Claude Shannon in the 1940s. From another, the idea has so many different shades of meaning that we shouldn’t be surprised (that’s a joke you will get in a bit) that it can be hard to understand.

Information theory is obviously an enormous subject, but we’re just giving thanks, not writing a textbook. I want to mention two ideas I find especially central. First, Shannon’s idea about relating information content to “surprisal.” Second, the very different intuitive notions of information that we get from engineering and physics.

Shannon, working at Bell Labs, was interested in the problem of how to send trustworthy signals efficiently over transatlantic cables. He was thinking about various ways to express information in a code: a set of symbols, each with a defined meaning. So a code might be an alphabet, or a set of words, or a literal cipher. And he noticed that there was a lot of redundancy in natural languages; the word “the” appears much more often in English than the word “axe,” although both have the same number of letters.

Let’s refer to each letter or symbol in a code as an “event.” Shannon’s insight was to realize that the more unlikely an event, the more information it conveyed when it was received. The statements “The Sun rose in the east this morning” and “The Sun rose in the west this morning” contain the same number of letters, but the former contains almost no information — you already were pretty sure the Sun would be rising in the east. But the latter, if obtained from a reliable source, would be very informative indeed, precisely because it was so unexpected. Clearly some kind of unprecedented astronomical catastrophe was in progress.

Imagine we can assign a probability p(x) to every different event x. Shannon wanted a way to quantify the information content of that event, which would satisfy various reasonable-seeming axioms: most crucially, that the information content of two independent events is the sum of the individual information contents. But the joint probability of two events is the product of their individual probabilities. So the natural thing to do would be to define the information content as the logarithm of the probability; the logarithm of a product equals the sum of the individual logarithms. But you want low probability to correspond to high information content, so Shannon defined the information content (also called the self-information, or surprisal, or Shannon information) of an event to be minus the log of the probability, which by math is equal to the log of the reciprocal of the probability:

    \[I(x) = - \log [p(x)] =\log \left(\frac{1}{p(x)}\right).\]

Note that probabilities are numbers between 0 and 1, and the log of such a number will be negative, with numbers closer to 0 being more negative than numbers closer to 1. So I(x) goes from +\infty at p(x)=0 to 0 at p(x)=1. An impossible message is infinitely surprising, and therefore conveys infinite information; an inevitable message is completely unsurprising, and conveys no information at all.

From there, Shannon suggested that we could characterize how efficient an entire code was at conveying information: just calculate the average (expectation value) of the information content for all possible events. When we have a probability distribution p(x), the average of any function f(x) is just the sum of the the values of the function times their respective probabilities, \langle f\rangle = \sum_x p(x) f(x). So we characterize the information content of a code via the quantity

    \[H[p] = - \sum_x p(x) \log[p(x)].\]

The only question is, what to call this lovely newly-defined quantity that surely nobody had ever thought of before? Happily Shannon was friends with John von Neumann, who informed him, “You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, no one really knows what entropy really is, so in a debate you will always have the advantage.” So entropy it is.

Indeed, this formula is precisely that which had been put forward (unknown to Shannon) by Josiah Willard Gibbs in the 1870’s as a definition of entropy in statistical mechanics. (It is related to the definition on Ludwig Boltzmann’s tombstone, S= k \log W, and Boltzmann had also suggested similar expressions to the above.) On the one hand, it seems remarkable to find precisely the same expression playing central roles in problems as disparate as sending signals across cables and watching cream mix into coffee; on the other hand, it’s a relatively simple expression and the axioms used to derive it are actually pretty similar, so perhaps we shouldn’t be surprised; on the third hand, the connection between information theory and statistical mechanics turns out to be deep and fruitful, so it’s more than just a mathematical coincidence.

But let me highlight the one aspect of the term “information” that can be sometimes confusing to people. To the engineer, a code that is maximally informative is one for which p(x) is relatively uniform over all events x, which means H[p(x)] is maximal or close to it; in that case, every event will tell you something at least a little bit interesting. For them, high entropy = high information.

But to a physicist who might be asking “how much information do I have about the state of a system?”, you have more information when p(x) is relatively narrowly concentrated around some value, rather than being all spread out. For them, high entropy = low information! Indeed, one physically-relevant notion of “information” is the “accessible information” of a system, which can be defined as H_\mathrm{max} - H. (I talk about this a bit in my recent solo podcast on complexity.)

Perhaps we shouldn’t be so surprised that physicists and engineers posit oppositely-directed relationships between entropy and information. It’s just a reflection of the fact that “information” is so ubiquitous and has so many different uses. We should be thankful that we’re beginning to understand it so well.

November 26, 2025

Tim GowersCreating a database of motivated proofs

It’s been over three years since my last post on this blog and I have sometimes been asked, understandably, whether the project I announced in my previous post was actually happening. The answer is yes — the grant I received from the Astera Institute has funded several PhD students and a couple of postdocs, and we have been busy. In my previous post I suggested that I would be open to remote collaboration, but that has happened much less, partly because a Polymath-style approach would have been difficult to manage while also ensuring that my PhD students would have work that they could call their own to put in their theses.

In general I don’t see a satisfactory solution to that problem, but in this post I want to mention a subproject of the main project that is very much intended to be a large public collaboration. A few months ago, a call came out from Renaissance Philanthropies saying that they were launching a $9m AI for Math Fund to spend on projects in the general sphere of AI and mathematics, and inviting proposals. One of the categories that they specifically mentioned was creating new databases, and my group submitted a proposal to create a database of what we call “structured motivated proofs,” a piece of terminology that I will explain a bit more later in just a moment. I am happy to report that our proposal was one of the 29 successful ones. Since a good outcome to the project will depend on collaboration from many people outside the group, we need to publicize it, which is precisely the purpose of this post. Below I will be more specific about the kind of help we are looking for.

Why might yet another database of theorems and proofs be useful?

The underlying thought behind this project is that AI for mathematics is being held back not so much by an insufficient quantity of data as by the wrong kind of data. (For a more general exploration of this theme, see here.) All mathematicians know, and some of us enjoy complaining about it, that it is common practice when presenting a proof in a mathematics paper, or even textbook, to hide the thought processes that led to the proof. Often this does not matter too much, because the thought processes may be standard ones that do not need to be spelt out to the intended audience. But when proofs start to get longer and more difficult, they can be hard to read because one has to absorb definitions and lemma statements that are not obviously useful, are presented as if they appeared from nowhere, and demonstrate their utility only much later in the argument.

A sign that this is a problem for AI is the behaviour one observes after asking an LLM to prove a statement that is too difficult for it. Very often, instead of admitting defeat, it will imitate the style of a typical mathematics paper and produce rabbits out of hats, together with arguments later on that those rabbits do the required job. The problem is that, unlike with a correct mathematics paper, one finds when one scrutinizes the arguments carefully that they are wrong. However, it is hard to find superficial features that distinguish between an incorrect rabbit with an incorrect argument justifying that rabbit (especially if the argument does not go into full detail) and a correct one, so the kinds of statistical methods used by LLMs do not have an easy way to penalize the incorrectness.

Of course, that does not mean that LLMs cannot do mathematics at all — they are remarkably good at it, at least compared with what I would have expected three years ago. How can that be, given the problem I have discussed in the previous paragraph?

The way I see it (which could change — things move so fast in this sphere), the data that is currently available to train LLMs and other systems is very suitable for a certain way of doing mathematics that I call guess and check. When trying to solve a maths problem, you will normally write down the routine parts of an argument without any fuss (and an LLM can do them too because it has seen plenty of similar examples), but if the problem as a whole is not routine, then at some point you have to stop and think, often because you need to construct an object that has certain properties (I mean this in a rather general way — the “object” might be a lemma that will split up the proof in a nice way) and it is not obvious how to do so. The guess-and-check approach to such moments is what it says: you make as intelligent a guess as you can and then see whether it has the properties you wanted. If it doesn’t, you make another guess, and you keep going until you get lucky.

The reason an LLM might be tempted to use this kind of approach is that the style of mathematical writing I described above makes it look as though that is what we as mathematicians do. Of course, we don’t actually do that, but we tend not to mention all the failed guesses we made and how we carefully examined why they failed, modifying them in appropriate ways in response, until we finally converged on an object that worked. We also don’t mention the reasoning that often takes place before we make the guess, saying to ourselves things like “Clearly an Abelian group can’t have that property, so I need to look for a non-Abelian group.”

Intelligent guess and check works well a lot of the time, particularly when carried out by an LLM that has seen many proofs of many theorems. I have often been surprised when I have asked an LLM a problem of the form \exists x\in X \ P(x), where P is some property that is hard to satisfy, and the LLM has had no trouble answering it. But somehow when this happens, the flavour of the answer given by the LLM leaves me with the impression that the technique it has used to construct x is one that it has seen before and regards as standard.

If the above picture of what LLMs can do is correct (the considerations for reinforcement-learning-based systems such as AlphaProof are not identical but I think that much of what I say in this post applies to them too for slightly different reasons), then the likely consequence is that if we pursue current approaches, then we will reach a plateau: broadly speaking they will be very good at answering a question if it is the kind of question that a mathematician with the right domain expertise and good instincts would find reasonably straightforward, but will struggle with anything that is not of that kind. In particular, they will struggle with research-level problems, which are, almost by definition, problems that experts in the area do not find straightforward. (Of course, there would probably be cases where an LLM spots relatively easy arguments that the experts had missed, but that wouldn’t fundamentally alter the fact that they weren’t really capable of doing research-level mathematics.)

But what if we had a database of theorems and proofs that did not hide the thought processes that lay behind the non-obvious details of the proofs? If we could train AI on a database of accounts of proof discoveries and if, having done so, we then asked it to provide similar accounts, then it would no longer resort to guess-and-check when it got stuck, because the proof-discovery accounts it had been trained on would not be resorting to it. There could be a problem getting it to unlearn its bad habits, but I don’t think that difficulty would be impossible to surmount.

The next question is what such a database might look like. One could just invite people to send in stream-of-consciousness accounts of how they themselves found certain proofs, but that option is unsatisfactory for several reasons.

  1. It can be very hard to remember where an idea came from, even a few seconds after one has had it — in that respect it is like a dream, the memory of which becomes rapidly less vivid as one wakes up.
  2. Often an idea will seem fairly obvious to one person but not to another.
  3. The phrase “motivated proof” means different things to different people, so without a lot of careful moderation and curation of entries, there is a risk that a database would be disorganized and not much more helpful than a database of conventionally written proofs.
  4. A stream-of-consciousness account could end up being a bit too much about the person who finds the proof and not enough about the mathematical reasons for the proof being feasibly discoverable.

To deal with these kinds of difficulties, we plan to introduce a notion of a structured motivated proof, by which we mean a proof that is generated in a very particular way that I will partially describe below. A major part of the project, and part of the reason we needed funding for it, is to create a platform that will make it convenient to input structured motivated proofs and difficult to insert the kinds of rabbits out of hats that make a proof mysterious and unmotivated. In this way we hope to gamify the task of creating the database, challenging people to input into our system proofs of certain theorems that appear to rely on “magic” ideas, and perhaps even offering prizes for proofs that contain steps that appear in advance to be particularly hard to motivate. (An example: the solution by Ellenberg and Gijswijt of the cap-set problem uses polynomials in a magic-seeming way. The idea of using polynomials came from an earlier paper of Croot, Lev and Pach that proved a closely related theorem, but in that paper it just appears in the statement of their Lemma 1, with no prior discussion apart from the words “in the present paper we use the polynomial method” in the introduction.)

What is a structured motivated proof?

I wrote about motivated proofs in my previous post, but thanks to many discussions with other members of the group, my ideas have developed quite a lot since then. Here are two ways we like to think about the concept.

1. A structured motivated proof is one that is generated by standard moves.

I will not go into full detail about what I mean by this, but will do so in a future post when we have created the platform that we would like people to use in order to input proofs into the database. But the basic idea is that at any one moment one is in a certain state, which we call a proof-discovery state, and there will be a set of possible moves that can take one from the current proof-discovery state to a new one.

A proof-discovery state is supposed to be a more formal representation of the state one is in when in the middle of solving a problem. Typically, if the problem is difficult, one will have asked a number of questions, and will be aware of logical relationships between them: for example, one might know that a positive answer to Q1 could be used to create a counterexample to Q2, or that Q3 is a special case of Q4, and so on. One will also have proved some results connected with the original question, and again these results will be related to each other and to the original problem in various ways that might be quite complicated: for example P1 might be a special case of Q2, which, if true would reduce Q3 to Q4, where Q3 is a generalization of the statement we are trying to prove.

Typically we will be focusing on one of the questions, and typically that question will take the form of some hypotheses and a target (the question being whether the hypotheses imply the target). One kind of move we might make is a standard logical move such as forwards or backwards reasoning: for example, if we have hypotheses of the form P(x) and \forall u\ P(u)\implies Q(u), then we might decide to deduce Q(x). But things get more interesting when we consider slightly less basic actions we might take. Here are three examples.

  1. We have in our list of hypotheses the fact that a function f is given by the formula f(x)=\exp(p(x)), where p is a polynomial, and our goal is to prove that there exists z such that f(z)=1. Without really thinking about it, we are conscious that f is a composition of two functions, one of which is continuous and one of which belongs to a class of functions that are all continuous, so f is continuous. Also, the conclusion \exists z\ f(z)=1 matches well the conclusion of the intermediate-value theorem. So the intermediate-value theorem comes naturally to mind and we add it to our list of available hypotheses. In practice we wouldn’t necessarily write it down, but the system we wish to develop is intended to model not just what we write down but also what is going on in our brains, so we propose a move that we call library extraction (closely related to what is often called premise selection in the literature). Note that we have to be a bit careful about library extraction. We don’t want the system to be allowed to call up results from the library that appear to be irrelevant but then magically turn out to be helpful, since those would feel like rabbits out of hats. So we want to allow extraction of results only if they are obvious given the context. It is not easy to define what “obvious” means, but there is a good rule of thumb for it: a library extraction is obvious if it is one of the first things ChatGPT thinks of when given a suitable non-cheating prompt. For example, I gave it the prompt, “I have a function f from the reals to the reals and I want to prove that there exists some z such that f(z)=1. Can you suggest any results that might be helpful?” and the intermediate-value theorem was its second suggestion. (Note that I had not even told it that f was continuous, so I did not need to make that particular observation before coming up with the prompt.)
  2. We have a goal of the form \exists x\in X\ P(x). If this were a Lean proof state, the most common way to discharge a goal of this form would be to input a choice for x. That is, we would instantiate the existential quantifier with some x_0 and our new goal would be P(x_0). However, as with library extraction, we have to be very careful about instantiation if we want our proof to be motivated, since we wish to disallow highly surprising choices of x_0 that can be found only after a long process of thought. So we have to restrict ourselves to obvious instantiations. One way that an instantiation in our system will count as obvious is if the variable is instantiated with a term that is already present in the proof-discovery state. If the desired term is not present, then in order to continue with the proof, it will be necessary to carry out moves that generate it. A very common technique for this is the use of metavariables: instead of guessing a suitable x_0, we create a variable x^\bullet and change the goal to P(x^\bullet), which we can think of as saying “I’m going to start trying to prove P(x^\bullet) even though I haven’t chosen x^\bullet yet. As the attempted proof proceeds, I will note down any properties Q_1,\dots,Q_k that x^\bullet might have that would help me finish the proof, in the hope that (i) I get to the end and (ii) the problem \exists x\ Q_1(x)\wedge\dots\wedge Q_k(x) is easier than the original problem.” Another kind of obvious instantiation is one where we try out an object that is “extreme” in some way — it might be the smallest element of X, or the largest, or the simplest. (Judging simplicity is another place where the ChatGPT rule of thumb can be used.)
  3. We cannot see how to answer the question we are focusing on so we ask a related question. Two very common kinds of related question (as emphasized by Polya) are generalization and specialization. Perhaps we don’t see why a hypothesis is helpful, so we see whether the result holds if we drop that hypothesis. If it does, then we are no longer distracted by an irrelevant hypothesis. If it does not, then we can hope to find a counterexample that will help us understand how to use the hypothesis. Or perhaps we are trying to prove a general statement but it is not clear how to do so, so instead we formulate some special cases, hoping that we can prove them and spot features of the proofs that we can generalize. Again we have to be rather careful here not to allow “non-obvious” generalizations and specializations. Roughly the idea there is that a generalization should be purely logical — for example, dropping a hypothesis is fine but replacing the hypothesis “f is twice differentiable” by “f is upper semicontinuous” is not — and that a specialization should be to a special case that counts as an obvious instantiation in the sense discussed just above.

2. A structured motivated proof is one that can be generated with the help of a point-and-click system.

This is a surprisingly useful way to conceive of what we are talking about, especially as it relates closely to what I was talking about earlier: imposing a standard form on motivated proofs (which is why we call them “structured” motivated proofs) and gamifying the process of producing them.

The idea is that a structured motivated proof is one that can be generated using an interface (which we are in the process of creating — at the moment we have a very basic prototype that has a few of the features we will need, but not yet the more interesting ones) that has one essential property: the user cannot type in data. So what can they do? They can select text that is on their screen (typically mathematical expressions or subexpressions), they can click buttons, choose items from drop-down menus, and accept or reject “obvious” suggestions made to them by the interface.

If, for example, the current goal is an existential statement \exists x\ P(x), then typing in a formula that defines a suitable x is not possible, so instead one must select text or generate new text by clicking buttons, choosing from short drop-down menus, and so on. This forces the user to generate x, which is our proxy for showing where the idea of using x came from.

Broadly speaking, the way the prototype works is to get an LLM to read a JSON object that describes the variables, hypotheses and goals involved in the proof state in a structured format, and to describe (by means of a fairly long prompt) the various moves it might be called upon to do. Thus, the proofs generated by the system are not formally verified, but that is not an issue that concerns us in practice since there will be a human in the loop throughout to catch any mistakes that the LLM might make, and this flexibility may even work to our advantage to better capture the fluidity of natural-language mathematics.

There is obviously a lot more to say about what the proof-generating moves are, or (approximately equivalently) what the options provided by a point-and-click system will be. I plan to discuss that in much more detail when we are closer to having an interface ready, the target for which is the end of this calendar year. But the aim of the project is to create a database of examples of proofs that have been successfully generated using the interface, which can then be used to train AI to play the generate-structured-motivated-proof game.

How to get involved.

There are several tasks that will need doing once the project gets properly under way. Here are some of the likely ones.

  1. The most important is for people to submit structured motivated (or move-generated) proofs to us on the platform we provide. We hope that the database will end up containing proofs of a wide range of difficulty (of two kinds — there might be fairly easy arguments that are hard to motivate and there might be arguments that are harder to follow but easier to motivate) and also a wide range of areas of mathematics. Our initial target, which is quite ambitious, is to have around 1000 entries by two years from now. While we are not in a position to accept entries yet, if you are interested in participating, then it is not too early to start thinking in a less formal way about how to convert some of your favourite proofs into motivated versions, since that will undoubtedly make it easier to get them accepted by our platform when it is ready.
  2. We are in the process of designing the platform. As I mentioned earlier, we already have a prototype, but there are many moves we will need it to be able to do that it cannot currently do. For example, the current prototype allows just a single proof state, which consists of some variable declarations, hypotheses, and goals. It does not yet support creating subsidiary proof states (which we would need if we wanted to allow the user to consider generalizations and specializations, for example). Also, for the moment the prototype gets an LLM to implement all moves, but some of the moves, such as applying modus ponens, are extremely mechanical and would be better done using a conventional program. (On the other hand, moves such as “obvious library extraction” or “provide the simplest example” are better done by an LLM.) Thirdly, a technical problem is that LaTeX is currently rendered as images, which makes it hard to select subexpressions, something we will need to be able to do in a non-clunky way. And the public version of the platform will need to be web-based and very convenient to use. We will want features such as being able to zoom out and look at some kind of dependency diagram of all the statements and questions currently in play, and then zoom in on various nodes if the user wishes to work on them. If you think you may be able (and willing) to help with some of these aspects of the platform, then we would be very happy to hear from you. For some, it would probably help to have a familiarity with proof assistants, while for others we would be looking for somebody with software engineering experience. The grant from the AI for Math Fund will allow us to pay for some of this help, at rates to be negotiated. We are not yet ready to specify in detail what help we need, but would welcome any initial expressions of interest.
  3. Once the platform is ready and people start to submit proofs, it is likely that, at least to start with, they will find that the platform does not always provide the moves they need. Perhaps they will have a very convincing account of where a non-obvious idea in the proof came from, but the system won’t be expressive enough for them to translate that account into a sequence of proof-generating moves. We will want to be able to react to such situations (if we agree that a new move is needed) by expanding the capacity of the platform. It will therefore be very helpful if people sign up to be beta-testers, so that we can try to get the platform to a reasonably stable state before opening it up to a wider public. Of course, to be a beta-tester you would need to have a few motivated proofs in mind.
  4. It is not obvious that every proof submitted via the platform, even if submitted successfully, would be a useful addition to the database. For instance, it might be such a routine argument that no idea really needs to have its origin explained. Or it might be that, despite our best efforts, somebody finds a way of sneaking in a rabbit while using only the moves that we have provided. (One way this could happen is if an LLM made a highly non-obvious suggestion that happened to work, in which case the rule of thumb that if an LLM thinks of it, it must be obvious, would have failed in that instance.) For this reason, we envisage having a team of moderators, who will check entries and make sure that they are good additions to the database. We hope that this will be an enjoyable task, but it may have its tedious aspects, so we envisage paying moderators — again, this expense was allowed for in our proposal to the AI for Math Fund.

If you think you might be interested in any of these roles, please feel free to get in touch. Probably the hardest recruitment task for us will be identifying the right people with the right mixture of mathematical knowledge and software engineering skills to help us turn the platform into a well-designed web-based one that is convenient and pleasurable to use. If you think you might be such a person, or if you have a good idea for how we should go about finding one, we would be particularly interested to hear from you.

In a future post, I will say more about the kinds of moves that our platform will allow, and will give examples of non-motivated proofs together with how motivated versions of those proofs can be found and entered using the platform (which may involve a certain amount of speculation about what the platform will end up looking like).

How does this relate to use of tactics in a proof assistant?

In one way, our “moves” can be regarded as tactics of a kind. However, some of the moves we will need are difficult to implement in conventional proof assistants such as Lean. In parallel with the work described above, we hope to create an interface to Lean that would allow one to carry out proof-discovery moves of the kind discussed above but with the proof-discovery states being collections of Lean proof states. Members of my group have already been working on this and have made some very interesting progress, but there is some way to go. However, we hope that at some point (and this is also part of the project pitched to the AI for Math Fund) we will have created another interface that will have Lean working in the background, so that it will be possible to generate motivated proofs that will be (or perhaps it is better to say include) proofs in Lean at the same time.

Another possibility that we are also considering is to use the output of the first platform (which, as mentioned above, will be fairly formal, but not in the strict sense of a language such as Lean) to create a kind of blueprint that can then be autoformalized automatically. Then we would have a platform that would in principle allow mathematicians to search for proofs while working on their computers without having to learn a formal language, with their thoughts being formalized as they go.

November 25, 2025

Jordan EllenbergI wrote about John von Neumann for the Free Press

They asked me to do this piece as part of their series on great American immigrants. (Sorry, but it looks like this is paywalled for non-subscribers; I can’t see it, at any rate.)

It was fun to write, especially to get a chance to push back on the idea of von Neumann as a space alien who humans could barely comprehend. The real guy was fully human and much more interesting. There was a part I wanted to put in about how von Neumann’s late essay “Can We Survive Technology?” has a lot in common with the introduction to the Communist Manifesto, despite von Neumann’s vigorous anti-Communism. But I ran out of room and time. Maybe later!

David Hoggsubstellar objects (brown dwarfs)

I spent the day at the NSBP / NSHP meeting in San José. My favorite session of the day was the morning astro session, which was entirely about brown dwarfs. I learned a lot in a very short time. Caprice Phillips (UCSC) introduced the session with an introduction to the scientific and technical questions in play. She put a lot of emphasis on using binaries and clusters to put detailed abundance ratios onto substellar objects. This was what I expected: I thought (walking in to this session) that all known abundance ratios for brown dwarfs were from such kinds of studies. I learned different (keep reading).

Gabriel Munoz Zarazua (SFSU) followed by showing spectra from M-dwarfs, brown dwarfs, and Jupiter. It definitely looks like a sequence. He does spectral fitting (what they call, in this business, retrievals). It looks like he is getting very good, somewhat precise, abundance ratios for the photospheres of substellar objects! I asked more about this in the question period, and apparently I am way behind the times (Emily Rauscher, Michigan, helpfully pointed this out to me): Now brown-dwarf photosphere models are so good, they can be used to measure abundances, and pretty well.

I also learned in this session (maybe from Jorge Sanchez, ASU, or maybe from Efrain Alvarado, SFSU) that there is a very strong mass–abundance relation in the Solar System. That is, we don't expect, if brown dwarfs form the way planets do, that the detailed abundances of the brown dwarfs will match exactly the detailed abundances of the primary stars. But now we are really in a position to test that. Sanchez showed that we can get, from even photometry, abundances for substellar objects in the Milky Way halo. Again, totally new to me! And he finds metallicities at or below −3. Alvarado showed data on an amazing system J1416, which is an L–T binary with no stellar companion. Apparently it is the only known completely substellar binary.

Tommaso DorigoBaby Steps In The Reinforcement Learning World

I am moving some baby steps in the direction of Reinforcement Learning (RL) these days. In machine learning, RL is a well-established and very promising avenue for the development of artificial intelligence, and the field is in rapid development. Unfortunately I have been left behind, as I never really needed to fiddle with those techniques for my research. Until recently.

read more

Jordan EllenbergSphere packing, cap set, and the Turán problem are all the same thing

OK, they are not really the same thing. But I got you reading, right?

Here’s the sense in which they’re all the same thing. Let G be a group and H < G a subgroup. Let m > 1 be an integer. Write Orb_m for the set of orbits of G by simultaneous left multiplication on (G/H)^m, and f for the natural map from (G/H)^m to Orb_m. Let R be a subset of Orb. We say a subset S of (G.H) is an R-set if f(S^m) lies in R.

Master question: How large can an R-set be?

A lot of natural and popular questions are of this form, and I would argue that the unpopular questions of this form are also pretty natural! Perhaps this observation has been made before, but it’s new to me.

(Remark that you should only read if you’re pedantic: if R doesn’t contain m-tuples with repeated elements, then S^m can’t map to R, because S^m does have elements like that. So silently, in the examples below, I will mostly mean S choose m rather than S^m. I don’t want to commit to that forever, so if you really want it to be S^m, you can deal with that by appopriately adding some degenerate orbits to R.)

Some examples:

Sphere packing: G is AO_n, the affine orthogonal group or group of rigid motions, and H is O_n, so G/H is R^n. Take m=2. Then Orb is identified with the nonnegative real numbers, and f takes a pair of points to the distance between them. Let R be the interval [d,D]. Then an R-set is a set of points in R^n such that the distance between any two is at most D (i.e. we’re contained in some fixed big sphere) and at least d (the points don’t get too close together.)

(If having to choose an upper bound D annoys you, feel free to take G = O_n and H = O_{n-1}; then Orb is a finite interval [0,r] and you can just take R to be [d,r].)

Cap set: G is the group of affine linear transformations of F_p^n, and H is GL_n(F_p), so G/H = F_p^n. Take m=3 again. Orb is whatever it is, but one of the orbits is the set of (x,y,z) such that x,y,z are distinct and x-2y+z=0. Call this orbit, oh, I dunno, 3AP. Let R = Orb – 3AP. Then an R-set is a cap set, and indeed we want to know how big an R-set can be.

Turán problem: G is S_n and H is S_2 x S_{n-2}, so G/H is the set of unordered pairs in [n], or the set of edges in the complete graph K_n. m is arbitrary. Now let Γ be a graph with m edges. Any injection from v(Γ) to [n] gives you a point in (G/H)^m, and all such are carried to each other by the action of S_n. Call that orbit [Γ], and let R be Orb – [Γ]. Now a subset S of (G/H)^m is an m-edge graph on n vertices, and S is an R-set just when no m distinct edges form a copy of H. How large can such an S be? That is the Turán number ex(n,Γ), and the Turán problem is to say whatever we can about it.

A lot of problems can be written in this form. (Especially a lot of problems Erdős worked on.) The Heilbronn triangle problem. Problems about subsets of space with forbidden angles. Turán problems for hypergraphs. The Guan-Ramos conjecture and the related Erdős matching conjecture. Bounds for error correcting codes (here take G to be the semidirect product of F_2^n by S_n, and H to be S_n, and m=2, so that Orb is just the set of Hamming distances and the choice of R exactly allows you to exclude whatever distances you want on differences between codewords.) Families of r-dimensional vector spaces of k^n such that any two intersect transversely. The happy ending problem. (G = AGL_2, H = GL_2, m arbitrary, Orb_m = m-tuples of points in the plane up to affine linear equivalence, R = m-tuples not forming a convex polygon.)

Variants

There are many variants of the master question, which allow you to incorporate an even wider range of popular questions under its generous sheltering wings. For instance: you could impose a bound on the size of f(S^m) instead of asking f(S^m) to lie inside a fixed R. (Now you’ve got the Erdős distinct distance problem.) Or instead of a hard constraint you could ask which pairs (|S|, f^{-1}(R)) are possible; the original question asks for which |S| the two entries can be equal. (Now you’ve got the Erdős unit distance problem.) And what do we mean by |S|, anyway? When G is finite, so’s S, but when G is a Lie group (and yes, friends, I do mean either real or p-adic), one had better ask which conditions on R guarantee that S is finite. (Easy exercise: show that a subset of R^2 such that no three points form an angle of less than 0.0001 degrees is finite.) (Harder: the Erdos-Szekeres theorem that R-sets are finite when R is the set of non-convex m-gons that appears in the happy ending problem.) At any rate, there’s no need to insist that S be finite. Maybe “how big can S” be means “how big can its Hausdorff dimension be!” (Still pretty easy exercise: show that a subset of Q_p^* in which no three points form an equilateral triangle has dimension at most log 2 / log p.)

Or, in any one of these cases, you could ask about the chromatic number version of this question: how many pieces do I need to partition (G/H)^m into if each piece P_i has f(P_i) disjoint from R? For sphere packing (O(n) version), this asks: how many pieces of diameter 1 suffice to cover a large sphere? The (spherical) sphere packing number provides an obvious lower bound. This has got to be a known problem, right?

Families

This part is going to be a little more vague and not completely thought out, but I want to write it down so I remember it.

Let me comment that these problems typically come in families. I’m going to be kind of vague about what one should mean by that, because I’m not wholly sure. Take the Turan problem, for example. That’s really a list of problems, one for each n, or at least each sufficiently large n, given by data G_n, H_n, and R_n. But in a way they’re all the same, right? In particular, Orb_m doesn’t change with n (or at least it doesn’t for n large enough) As it happens I am very familiar with this vague notion of seuqences of things, each with an action of S_n, which are somehow all the same — they are functors from the category FI of finite sets with injections, which I’ve written a lot about over the years. In particular, (G_n/H_n) is a finitely generated FI-set, which you can read more about in the paper of Speyer, Ramos, and White. Never mind the definitions; just accept that there’s a reasonable notion of what counts as a family of (G,H,R) instances in this context.

Similarly, I think the cap set problems as n varies should be thought of as a family.

What about the continuous cases? I’m a little less clear there but let me make a stab at at least one kind of thing that should count. Let G be the affine generalized orthogonal group on R^2, i.e. the group of rigid motions where you’re also allowed to dilate, and let H be GO(2). Then G/H is R^2 and Orb_3 is the set of similarity classes of triangles. Write K_n in Orb_3 for the set of triangles such that the ratio between two edges never exceeds n. Then a K_n-set has size bounded by const*n^2 (because a K_n set is more or less the same thing as a packing of unit circles in a circle of radius n.) If U is some OTHER set subset of orb, I would consider a family of problems to be given by R = U intersect K_n, with n growing.

In the first two cases, write (G/H)_n for (G_n/H_n) and in the third case write (G/H)_n for K_n. Note that in all three cases we have an easy asymptotic formula for the maximal size of a ((G/H)_n)-set. Call this size S_n. Then here’s what I’d like to guess.

Vaguely stated guess: In any family of R, if we write m_R(n) for the maximal size of an R-set in the problem indexed by n, then the limit

γ_R = lim_n log(m_R(n) / S_n)

exists. This should be a basic invariant of R that measures how “restrictive” it is.

One could even be more aggressive and ask whether there are constants c_R, γ_R such that

m_R(n) = c_R |S_n|^γ_R + o(|S_n|^γ_R).

which is true in lots of cases where the behavior of m_R(n) is understood!

Is that too abstract? Let me give some sense of what it means. For cap sets, it says the size of the largest cap set in F_q^n grows like q^{γn} or maybe q^{(γ+ε)n} or something. This is known and pretty easy, but what’s not at all known or easy is what γ is. I proved a theorem with Giswijt that γ < 1. I have no idea whether the more aggressive statement is true! I don’t know how to rule out that it’s some kind of funny q^{γ n + epsilon}. For the Turán problem, the guess says log ex(n,Γ) / log n approaches a limit in n; is that known?

If I asked how many points there could be in a unit line segment, no two separated by less than 1/n, and no “approximate three-term APs”: three points x,y,z satisfying y in [(1/2+δ)x + (1/2-δ)z, (1/2-δ)x + (1/2+δ)z], is that maximum asymptotic to some c n^γ, with γ depending on δ? In this context, one might imagine that γ is the largest dimension a set with no δ-approximate 3-term APs could have (throwing out the minimum distance requirement that forces the set to be finite.)

And the aggressive guess would say that, I dunno, if I asked how large a collection of k-element subsets of [n] could be if no m of them shared at least t points, the answer would also be asymptotic to some c n^γ. Believable! I have no reason to believe any of this except that the cases where these problems are solved, e.g. the problems discussed in this survey of set intersection problems, all seem to have answers of this form.

My instinct is that γ_R should be much easier to compute, and that c_R, which is more like a sphere-packing constant, should be more subtle.

Here’s a very concrete question. Take G = S_n, H = S_{n-k}, m = 2. So Orb_2 is just the set of double cosets H \ G / H, which indeed stabilizes for large n to some finite set you can describe combinatorially. For each subset R of this finite set, the guess says there’s some corresponding γ_R, and these numbers are monotone increasing under set inclusion. So… what are they?

I have not thought about this very hard! I am posting with the idea that people will tell me which parts of this are already understood, and which parts are wrong, and in what directions there’s more to say or ask.

Terence TaoCall for industry sponsors for IPAM’s RIPS program

Over the last 25 years, the Institute for pure and applied mathematics (IPAM) at UCLA (where I am now director of special projects) has run the popular Research in Industrial Projects for Students (RIPS) program every summer, in which industry sponsors with research projects are matched with talented undergraduates and a postdoctoral mentor to work at IPAM on the sponsored project for nine weeks. IPAM is now putting out a call for industry sponsors who can contribute a suitable research project for the Summer of 2026, as well as funding to cover the costs of the research; details are available here.

(Student applications for this program will open at a later date, once the list of projects is finalized.)

November 24, 2025

Terence TaoClimbing the cosmic distance ladder: another sample chapter

Five years ago, I announced a popular science book project with Tanya Klowden on the cosmic distance ladder, in which we released a sample draft chapter of the book, covering the “fourth rung” of the ladder, which for us meant the distances to the planets. In the intervening time, a number of unexpected events have slowed down this project significantly; but I am happy to announce that we have completed a second draft chapter, this time on the “seventh rung” of measuring distances across the Milky Way, which required the maturation of the technologies of photography and spectroscopy, as well as the dawn of the era of “big data” in the early twentieth century, as exemplified for instance by the “Harvard computers“.

We welcome feedback of course, and are continuing to work to complete the book despite the various delays. In the mean time, you can check out our instagram account for the project, or the pair of videos that Grant Sanderson (3blue1brown) produced with us on this topic, which I previously blogged about here.

Thanks to Clio Cresswell and Noah Klowden for comments and corrections.

November 23, 2025

Scott Aaronson Podcasts!

A 9-year-old named Kai (“The Quantum Kid”) and his mother interviewed me about closed timelike curves, wormholes, Deutsch’s resolution of the Grandfather Paradox, and the implications of time travel for computational complexity:

This is actually one of my better podcasts (and only 24 minutes long), so check it out!


Here’s a podcast I did a few months ago with “632nm” about P versus NP and my other usual topics:


For those who still can’t get enough, here’s an interview about AI alignment for the “Hidden Layers” podcast that I did a year ago, and that I think I forgot to share on this blog at the time:


What else is in the back-catalog? Ah yes: the BBC interviewed me about quantum computing for a segment on Moore’s Law.


As you may have heard, Steven Pinker recently wrote a fantastic popular book about the concept of common knowledge, entitled When Everyone Knows That Everyone Knows… Steve’s efforts render largely obsolete my 2015 blog post Common Knowledge and Aumann’s Agreement Theorem, one of the most popular posts in this blog’s history. But I’m willing to live with that, not only because Steven Pinker is Steven Pinker, but also because he used my post as a central source for the topic. Indeed, you should watch his podcast with Richard Hanania, where Steve lucidly explains Aumann’s Agreement Theorem, noting how he first learned about it from this blog.

November 22, 2025

n-Category Café Beyond the Geometry of Music

Yesterday I had a great conversation with Dmitri Tymoczko about groupoids in music theory. But at this Higgs Centre Colloquium, he preferred to downplay groupoids and talk in a way physicists would enjoy more. Click here to watch his talk!

What’s great is that Tymoczkyo not faking it: he’s really found deep ways in which symmetry shows up pervasively in music.

At first he tried to describe them geometrically using orbifolds, which are spaces in which some singular points have nontrivial symmetry groups, like the tip of a cone formed by modding out the plane by the action of the group /n\mathbb{Z}/n. But then he realized that the geometry was less important than the symmetry, which you can describe using groupoids. That’s why his talk is called “Beyond the geometry of music”.

I’m helping him with his work on groupoids, and I hope he explains his work to mathematicians someday without pulling his punches. I didn’t get to interview him yesterday, but I’ll try to do that soon.

For now you can read his books A Geometry of Music and Harmony: an Owner’s Manual along with many papers. What I’ve read so far is really exciting.

John BaezBeyond the Geometry of Music

Yesterday I had a great conversation with Dmitri Tymoczko about groupoids in music theory. But at this Higgs Centre Colloquium, he preferred to downplay groupoids and talk in a way physicists would enjoy more. Click on the picture to watch his talk!

What’s great is that he’s not faking it: he’s really found deep ways in which symmetry shows up pervasively in music.

At first he tried to describe them geometrically using ‘orbifolds’, which are spaces in which some singular points have nontrivial symmetry groups, like the tip of a cone. But then he realized that the geometry was less important than the symmetry, which you can describe using ‘groupoids’: categories where every morphism is invertible. That’s why his talk is called “Beyond the geometry of music”.

I’m helping him with his work on groupoids, and I hope he explains his work to mathematicians someday without pulling his punches. I didn’t get to interview him yesterday, but I’ll try to do that soon.

For now you can read his books A Geometry of Music and Harmony: an Owner’s Manual along with many papers. What I’ve read so far is really exciting.

November 21, 2025

Doug NatelsonQuantum geometry - some intuition

There has been a great growing interest in quantum geometry in recent years.  Last week, I heard an excellent talk by Raquel Queiroz about this that gave me a more physically intuitive interpretation  of this topic.  The more formal write-up is in this preprint from this past April, which I'd missed at the time.

Caution:  Math incoming.  I will try to give a more physical picture at the end.  I know that this won't be very readable to non-experts.    

As I've written before,  (e.g. here and a bit here), the electronic states in crystalline solids are often written as Bloch waves of the form \(u_{n\mathbf{k}}(\mathbf{r})\exp(i \mathbf{k}\cdot \mathbf{r})\), where \(u_{n\mathbf{k}}(\mathbf{r})\) is periodic in the spatial period of the crystal lattice.  For many years, the \(\mathbf{k}\) dependence of \(u_{n\mathbf{k}}(\mathbf{r})\) was comparatively neglected, but now it is broadly appreciated that this is the root of all kinds of interesting physics, including the anomalous Hall effect and its quantum version.  

We can compute how much \(u_{n\mathbf{k}}(\mathbf{r})\) changes with \(\mathbf{k}\).  The Berry connection is related to the phase angle racked up by moving around in \(\mathbf{k}\), and it's given by \( \mathbf{A}(\mathbf{k}) = i \langle u_{n\mathbf{k}}| \nabla_{\mathbf{k}}| u_{n\mathbf{k}} \rangle \).  One can define \(\mathbf{\Omega} \equiv \nabla \times \mathbf{A}(\mathbf{k})\) as the Berry curvature, and the "anomalous velocity" is given by \(-\dot{\mathbf{k}}\times \mathbf{\Omega}\).  

If we worry about possible changes in the magnitude as well, and \( |\langle u_{n\mathbf{k}}| u_{n\mathbf{k+dk}} \rangle |^{2} = 1 - g^{n}_{\mu \nu}dk_{\mu}dk_{\nu}\) plus higher order terms.  The quantity \(g^{n}_{\mu \nu}\) is the quantum metric, and it can be written in terms of dipole operators:  \(g^{n}_{\mu \nu}= \sum_{m\ne n}\langle u_{n,\mathbf{k}}|\hat{r}_{\mu}|u_{m \mathbf{k}}\rangle \langle u_{m,\mathbf{k}}|\hat{r}_{\nu}|u_{n \mathbf{k}}\rangle\).  The quantum metric quantifies the "distance between" the Bloch states as one moves around in \(\mathbf{k}\).  

That last bit is what I really learned from the talk.  Basically, if you try to consider electrons localized to a particular lattice site in real space, this can require figuring in states in multiple bands, and the matrix elements involve dipole operators.  The quantum geometric tensor \(g_{\mu \nu}\) quantifies the dipole fluctuations in the electronic density.  You can define a lengthscale \(\ell_{g}\equiv \sqrt{\mathrm{Tr} g}\), and this can tell you about the spatial scale of polarization fluctuations relative to, e.g., the lattice spacing.  Metals will have essentially divergent fluctuation lengthscales, while insulators have nicely bound charges (that give peaks in the optical conductivity at finite frequency).   The quantum geometry then influences all kinds of experimentally measurable quantities (see here).  

Neat stuff.  Someday I'd like to return to this with a nice cartoon/animation/presentation for non-experts.  The idea that there is so much richness within even relatively "boring" materials still amazes me.


Matt von HippelMandatory Dumb Acronyms

Sometimes, the world is silly for honest, happy reasons. And sometimes, it’s silly for reasons you never even considered.

Scientific projects often have acronyms, some of which are…clever, let’s say. Astronomers are famous for acronyms. Read this list, and you can find examples from 2D-FRUTTI and ABRACADABRA to WOMBAT and YORIC. Some of these aren’t even “really” acronyms, using letters other than the beginning of each word, multiple letters from a word, or both. (An egregious example from that list: VESTALE from “unVEil the darknesS of The gAlactic buLgE”.)

But here’s a pattern you’ve probably not noticed. I suggest that you should see more of these…clever…acronyms in projects in Europe, and they should show up in a wider range of fields, not just astronomy. And the reason why, is the European Research Council.

In the US, scientific grants are spread out among different government agencies. Typical grants are small, the kind of thing that lets a group share a postdoc every few years, with different types of grants covering projects of different scales.

The EU, instead, has the European Research Council, or ERC, with a flagship series of grants covering different career stages: Starting, Consolidator, and Advanced. Unlike most US grants, these are large (supporting multiple employees over several years), individual (awarded to a single principal investigator, not a collaboration) and general (the ERC uses the same framework across multiple fields, from physics to medicine to history).

That means there are a lot of medium-sized research projects in Europe that are funded by an ERC grant. And each of them are required to have an acronym.

Why? Who knows? “Acronym” is simply one of the un-skippable entries in the application forms, with a pre-set place of honor in their required grant proposal format. Nobody checks whether it’s a “real acronym”, so in practice it often isn’t, turning into some sort of catchy short name with “acronym vibes”. It, like everything else on these forms, is optimized to catch the attention of a committee of scientists who really would rather be doing something else, often discussed and refined by applicants’ mentors and sometimes even dedicated university staff.

So if you run into a scientist in Europe who proudly leads a group with a cutesy, vaguely acronym-adjacent name? And you keep running into these people?

It’s not a coincidence, and it’s not just scientists’ sense of humor. It’s the ERC.

Scott Aaronson Quantum Investment Bros: Have you no shame?

Near the end of my last post, I made a little offhand remark:

[G]iven the current staggering rate of hardware progress, I now think it’s a live possibility that we’ll have a fault-tolerant quantum computer running Shor’s algorithm before the next US presidential election. And I say that not only because of the possibility of the next US presidential election getting cancelled, or preempted by runaway superintelligence!

As I later clarified, I’ll consider this “live possibility” to be fulfilled even if a fault-tolerant Shor’s algorithm is “merely” used to factor 15 into 3×5—a milestone that seems a few steps, but only a few steps, away from what Google, Quantinuum, QuEra, and others have already demonstrated over the past year. After that milestone, I then expect “smooth sailing” to more and more logical qubits and gates and the factorization of larger and larger integers, however fast or slow that ramp-up proceeds (which of course I don’t know).

In any case, the main reason I made my remark was just to tee up the wisecrack about whether I’m not sure if there’ll be a 2028 US presidential election.


My remark, alas, then went viral on Twitter, with people posting countless takes like this:

A quantum expert skeptic who the bears quote all the time – Scott Aaronson – recently got very excited about a number of quantum advances. He now thinks there’s a possibility of running Shor before the next US president election – a timeline that lines up ONLY with $IONQ‘s roadmap, and NOBODY else’s! This represent a MAJOR capitulation of previously predicted timelines by any skeptics.

Shall we enumerate the layers of ugh here?

  1. I’ve been saying for several years now that anyone paranoid about cybersecurity should probably already be looking to migrate to quantum-resistant cryptography, because one can’t rule out the possibility that hardware progress will be fast. I didn’t “capitulate”: I mildly updated what I said before, in light of exciting recent advances.
  2. A “live possibility” is short not only of a “certainty,” but of a “probability.” It’s basically just an “I’m not confident this won’t happen.”
  3. Worst is the obsessive focus on IonQ, a company that I never mentioned (except in the context of its recently-acquired subsidiary, Oxford Ionics), but which now has a $17 billion valuation. I should explain that, at least since it decided to do an IPO, IonQ has generally been regarded within the research community as … err … a bit like the early D-Wave, intellectual-respectability-wise. They’ll eagerly sell retail investors on the use of quantum computers to recognize handwriting and suchlike, despite (I would say) virtually no basis to believe in a quantum scaling advantage for such tasks. Or they’ll aggressively market current devices to governments who don’t understand what they’re for, but just want to say they have a quantum computer and not get left behind. Or they’ll testify to Congress that quantum, unlike AI, “doesn’t hallucinate” and indeed is “deterministic.” It pains me to write this, as IonQ was founded by (and indeed, still employs) scientists who I deeply admire and respect.
  4. Perhaps none of this would matter (or would matter only to pointy-headed theorists like me) if IonQ were the world leader in quantum computing hardware, or even trapped-ion hardware. But by all accounts, IonQ’s hardware and demonstrations have lagged well behind those of its direct competitor, Quantinuum. It seems to me that, to whatever extent IonQ gets vastly more attention, it’s mostly just because it chose to IPO early, and also because it’s prioritized marketing to the degree it has.

Over the past few days, I’ve explained the above to various people, only to have them look back at me with glazed, uncomprehending eyes and say, “so then, which quantum stock should I buy? or should I short quantum?”

It would seem rude for me to press quarters into these people’s hands, explaining that they must make gain from whatever they learn. So instead I reply: “You do realize, don’t you, that I’m, like, a professor at a state university, who flies coach and lives in a nice but unremarkable house? If I had any skill at timing the market, picking winners, etc., don’t you think I’d live in a mansion with an infinity pool, and fly my Cessna to whichever conferences I deigned to attend?”


It’s like this: if you think quantum computers able to break 2048-bit cryptography within 3-5 years are a near-certainty, then I’d say your confidence is unwarranted. If you think such quantum computers, once built, will also quickly revolutionize optimization and machine learning and finance and countless other domains beyond quantum simulation and cryptanalysis—then I’d say that more likely than not, an unscrupulous person has lied to you about our current understanding of quantum algorithms.

On the other hand, if you think Bitcoin, and SSL, and all the other protocols based on Shor-breakable cryptography, are almost certainly safe for the next 5 years … then I submit that your confidence is also unwarranted. Your confidence might then be like most physicists’ confidence in 1938 that nuclear weapons were decades away, or like my own confidence in 2015 that an AI able to pass a reasonable Turing Test was decades away. It might merely be the confidence that “this still looks like the work of decades—unless someone were to gather together all the scientific building blocks that have now been demonstrated, and scale them up like a stark raving madman.” The trouble is that sometimes people, y’know, do that.

Beyond that, the question of “how many years?” doesn’t even interest me very much, except insofar as I can mine from it the things I value in life, like scientific understanding, humor, and irony.


There are, famously, many intellectual Communists who are ruthless capitalists in their day-to-day lives. I somehow wound up the opposite. Intellectually, I see capitalism as a golden goose, a miraculous engine that’s lifted the human species out of its disease-ridden hovels and into air-conditioned high-rises, whereas Communism led instead to misery and gulags and piles of skulls every single time it was tried.

And yet, when I actually see the workings of capitalism up close, I often want to retch. In case after case, it seems, our system rewards bold, confident, risk-taking ignoramuses and liars, those who can shamelessly hype a technology (or conversely, declare it flatly impossible)—with such voices drowning out the cautious experts who not only strive to tell the truth, but also made all the actual discoveries that the technology rests on. My ideal economic system is, basically, whichever one can keep the people who can clearly explain the capabilities and limits and risks and benefits of X in charge of X for as long as possible.

Jordan EllenbergMadison West JV girls basketball: Regents 45, LaFollette Lancers 37

AB is playing JV basketball this year, and tonight was her first home game. Lots of parents, and even more impressive, lots of fellow students. A lot of school spirit at West HS! The game was close for most of the going, with a lot of lead changes, and there was honestly a higher level of fan intensity than at some major league baseball games I’ve been to. (Looking at you, well-heeled SF Giants fans!) There is something very satisfying about yelling “DEE-FENSE” and stomping my feet and it is not a thing I get to do that often.

November 20, 2025

Terence TaoSum-difference exponents for boundedly many slopes, and rational complexity

I have uploaded to the arXiv my paper “Sum-difference exponents for boundedly many slopes, and rational complexity“. This is the second spinoff of my previous project with Bogdan Georgiev, Javier Gómez–Serrano, and Adam Zsolt Wagner that I recently posted about. One of the many problems we experimented using the AlphaEvolve tool with was that of computing sum-difference constants. While AlphaEvolve did modest improve one of the known lower bounds on sum-difference constants, it also revealed an asymptotic behavior to these constants that had not been previously observed, which I then gave a rigorous demonstration of in this paper.

In the original formulation of the sum-difference problem, one is given a finite subset {E} of {{\bf R}^2} with some control on projections, such as

\displaystyle  |\{ a: (a,b) \in E \}| \leq N

\displaystyle  |\{ b: (a,b) \in E \}| \leq N

\displaystyle  |\{ a+b: (a,b) \in E \}| \leq N

and one then asks to obtain upper bounds on the quantity

\displaystyle  |\{ a-b: (a,b) \in E \}|. \ \ \ \ \ (1)

This is related to Kakeya sets because if one joins a line segment between {(a,0)} and {(b,1)} for every {(a,b) \in E}, one gets a family of line segments whose set of directions has cardinality (1), but whose slices at heights {0,1,1/2} have cardinality at most {N}.

Because {a-b} is clearly determined by {a} and {b}, one can trivially get an upper bound of {N^2} on (1). In 1999, Bourgain utilized what was then the very recent “Balog–Szemerédi–Gowers lemma” to improve this bound to {N^{2-\frac{1}{13}}}, which gave a new lower bound of {\frac{13d+12}{25}} on the (Minkowski) dimension of Kakeya sets in {{\bf R}^d}, which improved upon the previous bounds of Tom Wolff in high dimensions. (A side note: Bourgain challenged Tom to also obtain a result of this form, but when they compared notes, Tom obtained the slightly weaker bound of {N^{2-\frac{1}{14}}}, which gave Jean great satisfaction.) Currently, the best upper bound known for this quantity is {N^{2-\frac{1}{6}}}.

One can get better bounds by adding more projections. For instance, if one also assumes

\displaystyle  |\{ a+2b: (a,b) \in E \}| \leq N

then one can improve the upper bound for (1) to {N^{2-\frac{1}{4}}}. The arithmetic Kakeya conjecture asserts that, by adding enough projections, one can get the exponent arbitrarily close to {1}. If one could achieve this, this would imply the Kakeya conjecture in all dimensions. Unfortunately, even with arbitrarily many projections, the best exponent we can reach asymptotically is {1.67513\dots}.

It was observed by Ruzsa that all of these questions can be equivalently formulated in terms of Shannon entropy. For instance, the upper bound {N^{2-\frac{1}{6}}} of (1) turns out to be equivalent to the entropy inequality

\displaystyle  {\bf H}(X-Y) \leq (2-\frac{1}{6}) \max( {\bf H}(X), {\bf H}(Y) , {\bf H}(X+Y) )

holding for all discrete random variables {X, Y} (not necessarily independent) taking values in {{\bf R}^2}. In the language of this paper, we write this as

\displaystyle  SD(\{0,1,\infty\}; -1) \leq 2-\frac{1}{6}.

Similarly we have

\displaystyle  SD(\{0,1,2,\infty\}; -1) \leq 2-\frac{1}{4}.

As part of the AlphaEvolve experiments, we directed this tool to obtain lower bounds for {SD(\{0,1,\infty\}; \frac{a}{b})} for various rational numbers {\frac{a}{b}}, defined as the best constant in the inequality

\displaystyle  {\bf H}(X+\frac{a}{b} Y) \leq SD(\{0,1,\infty\}; \frac{a}{b}) \max( {\bf H}(X), {\bf H}(Y) , {\bf H}(X+Y) ).

We did not figure out a way for AlphaEvolve to efficiently establish upper bounds on these quantities, so the bounds provided by AlphaEvolve were of unknown accuracy. Nevertheless, they were sufficient to give a strong indication that these constants decayed logarithmically to {2} as {a+b \rightarrow \infty}:

The first main result of this paper is to confirm that this is indeed the case, in that

\displaystyle  2 - \frac{c_2}{\log(2+|a|+|b|)} \leq SD(\{0,1,\infty\}; \frac{a}{b}) \leq 2 - \frac{c_1}{\log(2+|a|+|b|)}

whenever {a/b} is in lowest terms and not equal to {0, 1, \infty}, where {c_1,c_2>0} are absolute constants. The lower bound was obtained by observing the shape of the examples produced by AlphaEvolve, which resembled a discrete Gaussian on a certain lattice determined by {a,b}. The upper bound was established by an application of the “entropic Plünnecke–Ruzsa calculus”, relying particularly on the entropic Ruzsa triangle inequality, the entropic Balog–Szemerédi–Gowers lemma, as well as an entropy form of an inequality of Bukh.

The arguments also apply to settings where there are more projections under control than just the {0,1,\infty} projections. If one also controls projections {X + r_i Y} for various rationals {r_1,\dots,r_k} and {R} denotes the set of slopes of the projections under control, then it turns out that the associated sum-difference constant {SD(\{0,1,\infty,r_1,\dots,r_k\}; s)} still decays to {2}, but now the key parameter is not the height {|a|+|b|} of {s}, but rather what I call the rational complexity of {s} with respect to {R}, defined as the smallest integer {D} for which one can write {s} as a ratio {P(r_1,\dots,r_k)/Q(r_1,\dots,r_k)} where {P,Q} are integer-coefficient polynomials of degree at most {D} and coefficients at most {2^D}. Specifically, {SD(\{0,1,\infty,r_1,\dots,r_k\}; s)} decays to {2} at a polynomial rate in {D}, although I was not able to pin down the exponent of this decay exactly. The concept of rational complexity may seem somewhat artificial, but it roughly speaking measures how difficult it is to use the entropic Plünnecke–Ruzsa calculus to pass from control of {X, Y, X+Y}, and {X+r_i Y} to control of {X+sY}.

While this work does not make noticeable advances towards the arithmetic Kakeya conjecture (we only consider regimes where the sum-difference constant is close to {2}, rather than close to {1}), it does highlight the fact that these constants are extremely arithmetic in nature, in that the influence of projections {X+r_iY} on {X+sY} is highly dependent on how efficiently one can represent {s} as a rational combination of the {r_i}.

November 19, 2025

John BaezSafeguarded AI (Part 2)

60 people, including a lot of category theorists, are meeting in Edinburgh for the £59 million UK project called Safeguarded AI. I talked about it before here.

The plan is to build software that will let you precisely specify systems of many kinds, which an AI will design, and verify that what the AI designed meets your specifications. So: it’s not about building an AI, but instead, building a way to specify jobs for it and verify that it did those jobs correctly!

The director of this project, David Dalrymple, has changed the plan recently. There were many teams of category theorists designing formalisms to get this job done. David Jaz Myers at Topos Research UK was supposed to integrate all these formalisms. That would be a huge job.

But recently all but a few teams have been cut off from the main project—they can now do whatever they want. The project will focus on 3 parts:

1) The “categorical core”: a software infrastructure that lets you program using category theory concepts. I think Amar Hadzihasanovic, my former student Owen Lynch, and two others will be building this.

2) “DOTS”: the double operadic theory of systems, a general framework for building systems out of smaller parts. This is David Jaz Myers’ baby—see the videos.

3) Example applications. One of these, building colored Petri nets, will be done by my former student Jade Master. I don’t know all the others.

By September 2026, David Jaz Myers, Sophie Libkind, Matteo Capucci, Jason Brown and others are supposed to write a 300-page “thesis” on how this whole setup works. Some of the ideas are already available here:

• David Jaz Myers and Sophie Libkind, Towards a double operadic theory of systems.

It feels funny that so much of the math I helped invent is going into this project, and there’s a massive week-long meeting about it just a ten minute walk away, but I’m not involved. But this was by choice, and I’m happier just watching.

I apologize for any errors in the above, and for leaving out many other names of people who must be important in this project. I’ve spoken to various people involved, but not enough. I’m going to talk to David Jaz Myers tomorrow, but he wants to talk about what I’m really interested in these days: octonions and particle physics!

November 17, 2025

John BaezThe Inverse Cube Force Law

Newton’s Principia is famous for its investigations of the inverse square force law for gravity. But in this book Newton also did something that remained little-known until fairly recently. He figured out what kind of central force exerted upon a particle can rescale its angular velocity by a constant factor without affecting its radial motion. This turns out to be a force obeying an inverse cube law.

Given a particle in Euclidean space, a central force is a force that points toward or away from the origin and depends only on the particle’s distance from the origin. If the particle’s position at time t is \mathbf{r}(t) \in \mathbb{R}^n and its mass is some number m > 0, we have

m \, \ddot{\mathbf{r}}(t) = F(r(t))  \,\hat{\mathbf{r}}(t)

where \hat{\mathbf{r}}(t) is a unit vector pointing outward from the origin at the point \mathbf{r}(t). A particle obeying this equation always moves in a plane through the origin, so we can use polar coordinates and write the particle’s position as \bigl(r(t), \theta(t)\bigr). With some calculation one can show the particle’s distance from the origin, r(t), obeys

\displaystyle{ m \ddot r(t) = F(r(t)) + \frac{L^2}{mr(t)^3}  \qquad \qquad  \qquad \qquad (1) }

Here L = mr(t)^2 \dot \theta(t), the particle’s angular momentum, is constant in time. The second term in equation (1) says that the particle’s distance from the origin changes as if there were an additional force pushing it outward. This is a “fictitious force”, an artifact of working in polar coordinates. It is called the centrifugal force. And it obeys an inverse cube force law!

This explains Newton’s observation. Let us see why. Suppose that we have two particles moving in two different central forces F_1 and F_2, each obeying a version of equation (1), with the same mass m and the same radial motion r(t), but different angular momenta L_1 and L_2. Then we must have

\displaystyle{      F_1(r(t)) + \frac{L_1^2}{mr(t)^3}  =  F_2(r(t)) + \frac{L_2^2}{mr(t)^3}  }

If the particle’s angular velocities are proportional then L_2 = kL_1 for some constant k, so

\displaystyle{       F_2(r_1(t)) - F_1(r(t)) = \frac{(k^2 - 1)L_1^2}{mr(t)^3} }

This says that F_2 equals F_1 plus an additional inverse cube force.

A particle’s motion in an inverse cube force has curious features. First compare Newtonian gravity, which is an attractive inverse square force, say F(r) = -c/r^2 with c > 0. In this case we have

\displaystyle{ m \ddot r(t) = -\frac{c}{r(t)^2} + \frac{L^2}{mr(t)^3 } }

Because 1/r^3 grows faster than 1/r^2 as r \downarrow 0, as long as the angular momentum L is nonzero the repulsion of the centrifugal force will beat the attraction of gravity for sufficiently small r, and the particle will not fall in to the origin. The same is true for any attractive force F(r) = -c/r^p with p < 3. But an attractive inverse cube force can overcome the centrifugal force and make a particle fall in to the origin.

In fact there are three qualitatively different possibilities for the motion of a particle in an attractive inverse cube force F(r) = -c/r^3, depending on the value of c. With work we can solve for 1/r as a function of \theta (which is easier than solving for r). There are three cases depending on the value of

\displaystyle{  \omega^2 = 1 - \frac{cm}{L^2} }

vaguely analogous to the elliptical, parabolic and hyperbolic orbits of a particle in an inverse square force law:

\displaystyle{ \frac{1}{r(\theta)} } = \left\{ \begin{array}{lcl}   A \cos(\omega \theta) + B \sin(\omega \theta) & \text{if} & \omega^2 > 0 \\ \\   A + B \theta & \text{if} & \omega = 0 \\  \\   A e^{|\omega| \theta} + B e^{-|\omega| \theta}  & \text{if} & \omega^2 < 0   \end{array} \right.

The third case occurs when the attractive inverse cube force is strong enough to overcome the centrifugal force: c > L^2/m. Then the particle can spiral in to its doom, hitting the origin in a finite amount of time after infinitely many orbits, like this:

All three curves above are called Cotes spirals, after Roger Cotes’ work on the inverse cube force law, published posthumously in 1722. Cotes seems to have been the first to compute the derivative of the sine function. After Cotes’ death at the age of 33, Newton supposedly said “If he had lived we would have known something.”

The subtlety of the inverse cube force law is greatly heightened when we study it using quantum rather than classical mechanics. Here if c is too large the theory is ill-defined, because there is no reasonable choice of self-adjoint Hamiltonian. If c is smaller the theory is well-behaved. But at a certain borderline point it exhibits a remarkable property: spontaneous breaking of scaling symmetry. I hope to discuss this in my next column.

For more on the inverse cube force law, see:

• N. Grossman, The Sheer Joy of Celestial Mechanics, Birkhäuser, Basel, 1996, p. 34.

For more on Newton’s work involving the inverse cube force law, see:

• Wikipedia, Newton’s theorem of revolving orbits.

• S. Chandrasekhar, Newton’s Principia for the Common Reader, Oxford U. Press, Oxford, 1995, pp. 183–200.

Cotes’ book is

• Roger Cotes, Harmonia Mensuarum, Cambridge, 1722.

November 14, 2025

Scott Aaronson Quantum computing: too much to handle!

Tomorrow I’m headed to Berkeley for the Inkhaven blogging residency, whose participants need to write one blog post per day or get kicked out. I’ll be there to share my “wisdom” as a distinguished elder blogger (note that Shtetl-Optimized is now in its twentieth year). I’m acutely aware of the irony, that I myself can barely muster the willpower these days to put up a post every other week.

And it’s not as if nothing is happening in this blog’s traditional stomping-ground of quantum computing! In fact, the issue is just the opposite: way too much is happening for me to do it any sort of justice. Who do people think I am, Zvi Mowshowitz? The mere thought of being comprehensive, of responsibly staying on top of all the latest QC developments, makes me want to curl up in bed, and either scroll through political Substacks or take a nap.


But then, you know, eventually a post gets written. Let me give you some vignettes about what’s new in QC, any one of which could easily have been its own post if I were twenty years younger.

(1) Google announced verifiable quantum advantage based on Out-of-Time-Order-Correlators (OTOC)—this is actually from back in June, but it’s gotten more and more attention as Google has explained it more thoroughly. See especially this recent 2-page note by King, Kothari, et al., explaining Google’s experiment in theoretical computer science language. Basically, what they do is, starting from the all-|0⟩ state, to apply a random circuit C, then a single gate g, then C-1, then another gate h, then C again, then g again, then C-1, and then measure a qubit. If C is shallow, then the qubit is likely to still be |0⟩. If C is too deep, then the qubit is likely to be in the maximally mixed state, totally uncorrelated with its initial state—the gates g and h having caused a “butterfly effect” that completely ruined all the cancellation between C and C-1. Google claims that, empirically, there’s an intermediate regime where the qubit is neither |0⟩ nor the maximally mixed state, but a third thing—and that this third thing seems hard to determine classically, using tensor network algorithms or anything else they’ve thrown at it, but it can of course be determined by running the quantum computer. Crucially, because we’re just trying to estimate a few parameters here, rather than sample from a probability distribution (as with previous quantum supremacy experiments), the output can be checked by comparing it against the output of a second quantum computer, even though the problem still isn’t in NP. Incidentally, if you’re wondering why they go back and forth between C and C-1 multiple times rather than just once, it’s to be extra confident that there’s not a fast classical simulation. Of course there might turn out to be a fast classical simulation anyway, but if so, it will require a new idea: gauntlet thrown.

(2) Quantinuum, the trapped-ion QC startup in Colorado, announced its Helios processor. Quick summary of the specs: 98 qubits, all-to-all 2-qubit gates with 99.92% fidelity, the ability to choose which gates to apply “just in time” (rather than fixing the whole circuit in advance, as was needed with their previous API), and an “X”-shaped junction for routing qubits one way or the other (the sort of thing that a scalable trapped-ion quantum computer will need many of). This will enable, and is already enabling, more and better demonstrations of quantum advantage.

(3) Quantinuum and JP Morgan Chase announced the demonstration of a substantially improved version of my and Shih-Han-Hung’s protocol for generating cryptographically certified random bits, using quantum supremacy experiments based on random circuit sampling. They did their demo on Quantinuum’s new Helios processor. Compared to the previous demonstration, the new innovation is to send the circuit to the quantum computer one layer at a time, rather than all at once (something that, again, Quantinuum’s new API allows). The idea is that a cheating server, who wanted to spoof the randomness deterministically, now has much less time: using the most competitive known methods (e.g., those based on tensor network contraction), it seems the cheater would need to swing into action only after learning the final layer of gates, so would now have mere milliseconds to spoof rather than seconds, making Internet latency the dominant source of spoofing time in practice. While a complexity-theoretic analysis of the new protocol (or, in general, of “layer-by-layer” quantum supremacy protocols like it) is still lacking, I like the idea a lot.

(4) The startup company BlueQubit announced a candidate demonstration of verifiable quantum supremacy via obfuscated peaked random circuits, again on a Quantinuum trapped-ion processor (though not Helios). In so doing, BlueQubit is following the program that Yuxuan Zhang and I laid out last year: namely, generate a quantum circuit C that hopefully looks random to any efficient classical algorithm, but that conceals a secret high-probability output string x, which pops out if you run C on a quantum computer on the all-0 initial state. To try to hide x, BlueQubit uses at least three different circuit obfuscation techniques, which already tells you that they can’t have complete confidence in any one of them (since if they did, why the other two?). Nevertheless, I’m satisfied that they tried hard to break their own obfuscation, and failed. Now it’s other people’s turn to try.

(5) Deshpande, Fefferman, et al. announced a different theoretical proposal for quantum advantage from peaked quantum circuits, based on error-correcting codes. This seems tempting to try to demonstrate along the way to quantum fault-tolerance.

(6) A big one: John Bostanci, Jonas Haferkamp, Chinmay Nirkhe, and Mark Zhandry announced a proof of a classical oracle separation between the complexity classes QMA and QCMA, something that they’ve been working on for well over a year. Their candidate problem is basically a QMA-ified version of my Forrelation, which Raz and Tal previously used to achieve an oracle separation between BQP and PH. I caution that their paper is 91 pages long and hasn’t yet been vetted by independent experts, and there have been serious failed attempts on this exact problem in this past. If this stands, however, it finally settles a problem that’s been open since 2002 (and which I’ve worked on at various points starting in 2002), and shows a strong sense in which quantum proofs are more powerful than classical proofs. Note that in 2006, Greg Kuperberg and I gave a quantum oracle separation between QMA and QCMA—introducing the concept of quantum oracles for the specific purpose of that result—and since then, there’s been progress on making the oracle steadily “more classical,” but the oracle was always still randomized or “in-place” or had restrictions on how it could be queried.

(7) Oxford Ionics (which is now owned by IonQ) announced a 2-qubit gate with 99.99% fidelity: a record, and significantly past the threshold for quantum fault-tolerance. However, as far as I know, it remains to demonstrate this sort of fidelity in a large programmable system with dozens of qubits and hundreds of gates.

(8) Semi-announcement: Quanta reports that “Physicists Take the Imaginary Numbers Out of Quantum Mechanics,” and this seems to have gone viral on my social media. The article misses the opportunity to explain that “taking the imaginary numbers out” is as trivial as choosing to call each complex amplitude “just an ordered pair of reals, obeying such-and-such rules, which happen to mimic the rules for complex numbers.” Thus, the only interesting question here is whether one can take imaginary numbers out of QM in various more-or-less “natural” ways: a technical debate that the recent papers are pushing forward. For what it’s worth, I don’t expect that anything coming out of this line of work will ever be “natural” enough for me to stop explaining QM in terms of complex numbers in my undergraduate class, for example.

(9) The list of accepted talks for the annual QIP conference, to be held January 24-30 in Riga, Latvia, is now out. Lots of great stuff as always.

(10) There are probably other major recent developments in QC that I should’ve put into this post but forgot about. You can remind me about them in the comments.

(11) Indeed there are! I completely forgot that Phasecraft announced two simulations of fermionic systems that might achieve quantum advantage, one using Google’s Willow superconducting chip and the other using a Quantinuum device.


To summarize three takeaways:

  • Evidence continues to pile up that we are not living in the universe of Gil Kalai and the other quantum computing skeptics. Indeed, given the current staggering rate of hardware progress, I now think it’s a live possibility that we’ll have a fault-tolerant quantum computer running Shor’s algorithm before the next US presidential election. And I say that not only because of the possibility of the next US presidential election getting cancelled, or preempted by runaway superintelligence!
  • OK, but what will those quantum computers be useful for? Anyone who’s been reading this blog for the past 20 years, or any non-negligible fraction thereof, hopefully already has a calibrated sense of that, so I won’t belabor. But briefly: yes, our knowledge of useful quantum algorithms has slowly been expanding over the past thirty years. The central difficulty is that our knowledge of useful classical algorithms has also been expanding, and the only thing that matters is the differential between the two! I’d say that the two biggest known application areas for QC remain (a) quantum simulation and (b) the breaking of public-key cryptography, just as they were thirty years ago. In any case, none of the exciting developments that I’ve chosen to highlight in this post directly address the “what is it good for?” question, with the exception of the certified randomness thing.
  • In talks over the past three years, I’ve advocated “verifiable quantum supremacy on current hardware” as perhaps the central challenge right now for quantum computing theory. (As I love to point out, we do know how to achieve any two of (a) quantum supremacy that’s (b) verifiable and (c) runs on current hardware!) So I’m gratified that three of the recent developments that I chose to highlight, namely (1), (4), and (5), directly address this challenge. Of course, we’re not yet sure whether any of these three attempts will stand—that is, whether they’ll resist all attempts to simulate them classically. But the more serious shots on goal we have (and all three of these are quite serious), the better the chances that at least one will stand! So I’m glad that people are sticking their necks out, proposing these things, and honestly communicating what they know and don’t know about them: this is exactly what I’d hoped would happen. Of course, complexity-theoretic analysis of these proposals would also be great, perhaps from people with more youth and/or energy than me. Now it’s time for me to sleep.

Matt von HippelReminder to Physics Popularizers: “Discover” Is a Technical Term

When a word has both an everyday meaning and a technical meaning, it can cause no end of confusion.

I’ve written about this before using one of the most common examples, the word “model”, which means something quite different in the phrases “large language model”, “animal model for Alzheimer’s” and “model train”. And I’ve written about running into this kind of confusion at the beginning of my PhD, with the word “effective”.

But there is one example I see crop up again and again, even with otherwise skilled science communicators. It’s the word “discover”.

“Discover”, in physics, has a technical meaning. It’s a first-ever observation of something, with an associated standard of evidence. In this sense, the LHC discovered the Higgs boson in 2012, and LIGO discovered gravitational waves in 2015. And there are discoveries we can anticipate, like the cosmic neutrino background.

But of course, “discover” has a meaning in everyday English, too.

You probably think I’m going to say that “discover”, in everyday English, doesn’t have the same statistical standards it does in physics. That’s true of course, but it’s also pretty obvious, I don’t think it’s confusing anybody.

Rather, there is a much more important difference that physicists often forget: in everyday English, a discovery is a surprise.

“Discover”, a word arguably popularized by Columbus’s discovery of the Americas, is used pretty much exclusively to refer to learning about something you did not know about yet. It can be minor, like discovering a stick of gum you forgot, or dramatic, like discovering you’ve been transformed into a giant insect.

Now, as a scientist, you might say that everything that hasn’t yet been observed is unknown, ready for discovery. We didn’t know that the Higgs boson existed before the LHC, and we don’t know yet that there is a cosmic neutrino background.

But just because we don’t know something in a technical sense, doesn’t mean it’s surprising. And if something isn’t surprising at all, then in everyday, colloquial English, people don’t call it a discovery. You don’t “discover” that the store has milk today, even if they sometimes run out. You don’t “discover” that a movie is fun, if you went because you heard reviews claim it would be, even if the reviews might have been wrong. You don’t “discover” something you already expect.

At best, maybe you could “discover” something controversial. If you expect to find a lost city of gold, and everyone says you’re crazy, then fine, you can discover the lost city of gold. But if everyone agrees that there is probably a lost city of gold there? Then in everyday English, it would be very strange to say that you were the one who discovered it.

With this in mind, the way physicists use the word “discover” can cause a lot of confusion. It can make people think, as with gravitational waves, that a “discovery” is something totally new, that we weren’t pretty confident before LIGO that gravitational waves exist. And it can make people get jaded, and think physicists are overhyping, talking about “discovering” this or that particle physics fact because an experiment once again did exactly what it was expected to.

My recommendation? If you’re writing for the general public, use other words. The LHC “decisively detected” the Higgs boson. We expect to see “direct evidence” of the cosmic neutrino background. “Discover” has baggage, and should be used with care.

Matt Strassler Event with Professor Daniel Whiteson on Monday November 17 at 7pm

Next Monday, November 17th at 7pm, I’ll be at the Harvard Bookstore with particle physicist and author Daniel Whiteson. Professor Whiteson and his co-author Andy Warner have a nice new book, for the general science-aware reader, exploring an age-old and unanswered question: how universal is the knowledge and understanding that we call “physics”? How much of modern physics is actually telling us about the universe, and how much of it is created by, or an accident of, the humans who have helped bring it about?

For instance, if we started all over again and reran history from scratch, would the physics (and science more generally) of this re-run culture look much like our own, or might it turn out very differently? If another culture on Earth had had time to develop highly mature science (or something like it) in its own direction, independent of Western Europe’s influence, how different might that science be? (Indeed, would our word “science” even be translatable into their worldview?) Or if we encountered aliens with far greater understanding of the universe than we have, would we be able to recognize, parse, grok, appreciate, comprehend, and/or otherwise make sense of their notions of scientific knowledge?

Whiteson and his co-author, wanting to write a popular book rather than a scholarly one, and desiring nevertheless to take on these serious and challenging intellectual questions, have set their focus mostly on the aliens, accompanied by amusing cartoons and a generous helping of dad jokes (hey, some dad jokes are actually very funny.) They’re looking for a broad audience, and hopefully they will get it. But don’t let the light-hearted title (“Do Aliens Speak Physics?“) or the charmingly goofy cover fool you: this book might well make you laugh, but I guarantee it will make you think. Whether you’re just curious about science or you’ve been doing science yourself for years, I suspect that, within the vast array of problems and issues that are raised in this broad-minded book, there will be some you’ve never thought of.

Among scientists and philosophers, there are some who believe that any aliens with the capacity to reach the Earth will obviously “speak physics” — that math and physics float above contingencies of culture and species, and will easily be translated from any intelligent creature to any other. But are they perhaps flying too high? It’s clear that Whiteson and Warner are aiming to poke some holes — lots of holes —- in their hot-air balloon, and to do so in a way that a wide variety of readers can appreciate and enjoy.

I tend to agree with Whiteson on a lot of these issues, but that won’t stop me from asking him some tough questions. You can ask him some tough questions too, if you like — just come to the Harvard Bookstore at 7:00 on Monday and join the conversation!

November 12, 2025

Terence TaoNew Nikodym set constructions over finite fields

I have uploaded to the arXiv my paper “New Nikodym set constructions over finite fields“. This is a spinoff of my previous project with Bogdan Georgiev, Javier Gómez–Serrano, and Adam Zsolt Wagner that I recently posted about. In that project we experimented with using AlphaEvolve (and other tools, such as DeepThink and AlphaProof) to explore various mathematical problems which were connected somehow to an optimization problem. For one of these — the finite field Nikodym set problem — these experiments led (by a somewhat convoluted process) to an improved asymptotic construction of such sets, the details of which are written up (by myself rather than by AI tools) in this paper.

Let {{\mathbb F}_q} be a finite field of some order {q} (which must be a prime or a power of a prime), and let {d} be a fixed dimension. A Nikodym set in {{\mathbb F}_q^d} is a subset {N} of {{\mathbb F}_q^d} with the property that for every point {x \in {\mathbb F}_q^d}, there exists a line {\ell} passing through {x} such that all points of {\ell} other than {x} lie in {N}. Such sets are close cousins of Kakeya sets (which contain a line in every direction); indeed, roughly speaking, applying a random projective transformation to a Nikodym set will yield (most of) a Kakeya set. As a consequence, any lower bound on Kakeya sets implies a similar bound on Nikodym sets; in particular, one has a lower bound

\displaystyle  |N| \geq \frac{q^d}{2^{d-1}} + O(q^{d-1})

on the size of a Nikodym set {N}, coming from a similar bound on Kakeya sets due to Bukh and Chao using the polynomial method.

For Kakeya sets, Bukh and Chao showed this bound to be sharp up to the lower order error {O(q^{d-1})}; but for Nikodym sets it is conjectured that in fact such sets should asymptotically have full density, in the sense that

\displaystyle  |N| \geq q^d - o(q^d).

This is known in two dimensions thanks to work by Szönyi et al. on blocking sets, and was also established in bounded torsion cases (and in particular for even {q}) by Guo, Kopparty, and Sudan by combining the polynomial method with the theory of linear codes. But in other cases this conjecture remains open in three and higher dimensions.

In our experiments we focused on the opposite problem of constructing Nikodym sets of size as small as possible. In the plane {d=2}, constructions of size

\displaystyle  |N| = q^2 - q^{3/2} + O(q \log q) \ \ \ \ \ (1)

when {q} is a perfect square were constructed by Blokhuis et al, again using the theory of blocking sets; by taking Cartesian products of such sets, one can also make similar constructions in higher dimensions, again assuming {q} is a perfect square. Apart from this, though, there are few such constructions in the literature.

We set AlphaEvolve to try to optimize the three dimensional problem with a variable field size {q} (which we took to be prime for simplicity), with the intent to get this tool to come up with a construction that worked asymptotically for large {q}, rather than just for any fixed value of {q}. After some rounds of evolution, it arrived at a construction which empirically had size about {q^3 - 8q^2}. Inspecting the code, it turned out that AlphaEvolve had constructed a Nikodym set {N} by (mostly) removing eight low-degree algebraic surfaces (all of the form {\{ (x,y,x^i y)\}} for various {i}). We used the tool DeepThink to confirm the Nikodym property and to verify the construction, and then asked it to generalize the method. By removing many more than eight surfaces, and using some heuristic arguments based on the Chebotarev density theorem, DeepThink claimed a construction of size

\displaystyle  |N| = q^3 - 2 q^2 \log q + o(q^2 \log q) \ \ \ \ \ (2)

formed by removing several higher degree surfaces, but it acknowledged that the arguments were non-rigorous.

The arguments can be sketched here as follows. Let {V} be a random surface of degree {D}, and let {x} be a point in {{\mathbb F}_q^3} which does not lie in {V}. A random line through {x} then meets {V} in a number of points, which is basically the set of zeroes in {{\mathbb F}_q} of a random polynomial of degree {D}. The (function field analogue of the) Chebotarev density theorem predicts that the probability that this polynomial has no roots in {{\mathbb F}_q} is about {\delta_D}, where

\displaystyle  \delta_D = 1 - \frac{1}{1!} + \frac{1}{2!} - \dots + \frac{(-1)^D}{D!}

is the proportion of permutations on {D} elements that are derangements (no fixed points). So, if one removes {k} random surfaces of degrees {D_1,\dots,D_k}, the probability that a random line avoids all of these surfaces is about {\delta_{D_1} \dots \delta_{D_k}}. If this product is significantly greater than {1/q^2}, then the law of large numbers (and concentration of measure) then predicts (with high probability) that out of the {\sim q^2} lines through {x}, at least one will avoid the removed surfaces, thus giving (most of) a Nikodym set. The Lang-Weil estimate predicts that each surface has cardinality about {q^2}, so this should give a Nikodym set of size about {q^3 - kq^2}.

DeepThink took the degrees {D_1,\dots,D_k} to be large, so that the derangement probabilities {\delta_{D_i}} were close to {1/e}. This led it to predict that {k} could be taken to be as large as {2 \log q}, leading to the claimed bound (2). However, on inspecting this argument we realized that these moderately high degree surfaces were effectively acting as random sets, so one could dramatically simplify DeepThink’s argument by simply taking {N} to be a completely random set of the desired cardinality (2), in which case the verification of the Nikodym set property (with positive probability) could be established by a standard Chernoff bound-type argument (actually, I ended up using Bennett’s inequality rather than Chernoff’s inequality, but this is a minor technical detail).

On the other hand, the derangement probabilities {\delta_D} oscillate around {1/e}, and in fact are as large as {1/2} when {D=2}. This suggested that one could do better than the purely random construction if one only removed quadratic surfaces instead of higher degree surfaces, and heuristically predicted the improvement

\displaystyle  |N| = q^3 - \frac{2}{\log 2} q^2 \log q + o(q^2 \log q). \ \ \ \ \ (3)

However, our experiments with both AlphaEvolve and DeepThink to try to make this idea work either empirically, heuristically, or rigorously were all unsuccessful! Eventually Deepthink discovered the problem: random quadratic polynomials often had two or zero roots (depending on whether the discriminant was a non-zero quadratic residue, or a nonresidue), but would only very rarely have just one root (the discriminant would have to vanish). As a consequence, if {x} happened to lie on one of the removed quadratic surfaces {V}, it was extremely likely that most lines through {x} would intersect {V} in a further point; only the small minority of lines that were tangent to {V} and {x} would avoid this. None of the AI tools we tried were able to overcome this obstacle.

However, I realized that one could repair the construction by adding back a small random portion of the removed quadratic surfaces, to allow for a non-zero number of lines through {x} to stay inside the putative Nikodym set even when {x} was in one of the surfaces {V}, and the line was not tangent to {V}. Pursuing this idea, and performing various standard probabilistic calculations and projective changes of variable, the problem essentially reduced to the following: given {k} random quadratic polynomials in the plane {{\mathbb F}_q^2}, is it true that these polynomials simultaneously take quadratic residue values for {\gg 2^{-k}} of the points in that plane? Heuristically this should be true even for {2^{-k}} close to {1/q^2}. However, it proved difficult to accurately control this simultaneous quadratic residue event; standard algebraic geometry tools such as the Weil conjectures seemed to require some vanishing of étale cohomology groups in order to obtain adequate error terms, and this was not something I was eager to try to work out. However, by exploiting projective symmetry (and the {2}-transitive nature of the projective linear group), I could get satisfactory control of such intersections as long as {2^{-k}} was a little bit larger than {1/q} rather than {1/q^2}. This gave an intermediate construction of size

\displaystyle  |N| = q^3 - (\frac{1}{\log 2} + 1) q^2 \log q + o(q^2 \log q),

which still beat the purely random construction, but fell short of heuristic predictions. This argument (generalized to higher dimensions) is what is contained in the paper. I pose the question of locating a construction with the improved bound (3) (perhaps by some modification of the strategy of removing quadratic varieties) as an open question.

We also looked at the two-dimensional case to see how well AlphaEvolve could recover known results, in the case that {q} was a perfect square. It was able to come up with a construction that was slightly worse than the best known construction, in which one removed a large number of parabolas from the plane; after manually optimizing the construction we were able to recover the known bound (1). This final construction is somewhat similar to existing constructions (it has a strong resemblance to a standard construction formed by taking the complement of a Hermitian unital), but is still technically a new construction, so we have also added it to this paper.

November 10, 2025

John BaezThe Standard Model – Part 3

Physics is really bizarre and wonderful. Here I start explaining why the Standard Model has U(1) × SU(2) × SU(3) as its symmetry group. But I don’t assume you know anything about groups or quantum mechanics! So I have to start at the beginning: how the electromagnetic, weak, and strong force are connected to the numbers 1, 2, and 3. It’s all about quunits, qubits and qutrits.

You’ve heard of bits, which describe a binary alternative, like 0 and 1. You’ve probably heard about qubits, which are the quantum version of bits. The weak force is connected to qubits where the 2 choices are called “isospin up” and “isospin down”. The most familiar example is the choice between a proton and a neutron. A better example is the choice between an up quark and a down quark.

The strong force is connected to qutrits—the quantum version of a choice between 3 alternatives. In physics these are whimsically called “red”, “green” and “blue”. Quarks come in 3 colors like this.

The electromagnetic force is connected to “quunits” – the quantum version of a choice between just one alternative. It may seem like that’s no choice at all! But quantum mechanics is weird: there’s just one choice, but you can still rotate that choice.

Yes, I know this stuff sounds crazy. But this is how the world actually works. I start explaining it here, and I’ll keep on until it’s all laid out quite precisely.

November 09, 2025

Tommaso DorigoRestoring The Value Of Truth

Truth is under attack. It has always been, of course, because truth has always been a mortal enemy for those who attempt to seize or keep power in their hands. But the amplification of the phenomenon by today's information technology is extremely worrisome. AI today can generate fake videos and images that even experts have trouble flagging as such. This, combined with the different news value and propagation potential of false information with respect to typically less attention-grabbing true facts has created an explosive situation. What to do?

read more

November 08, 2025

Doug NatelsonVortices everywhere

The 2026 APS Oliver E. Buckley Prize in condensed matter physics was announced this week, and it's a really interesting combination of topics that, to a lay person, may seem to be completely unrelated.  

Fig. 1 from this follow-up PRB.
On the one hand, John Reppy (at age 94!) and Dave Bishop were honored for their work examining the properties of vortices in thin films of superfluid helium-4.  Relevant papers include this one from 1977, where they used a torsion pendulum coated with the helium film to examine the transition between normal and superfluid.  When the helium becomes a superfluid, it has (at low speeds) no viscosity, so it no longer has to rotate with the torsion pendulum; this means the rotational moment of inertia goes from that of (pendulum+helium) to just (pendulum), and the period of the oscillations increases.  Really detailed measurements of the oscillations and their damping allowed Reppy and Bishop to compare with models of the superfluid transition based on work by Kosterlitz and Thouless (and Berezinskii).  See the image for a diagram of the experimental setup - very clever and intricate.  

The key idea here is the role of vortices.  Superfluidity in helium is described by an order parameter that looks like a wavefunction - it has an amplitude, \(\Psi_{0}\), and a phase \(\phi\), so that \(\Psi(\mathbf{r}) = \Psi_{0} \exp(i \phi)\).   That order parameter is supposed to be single-valued, meaning if you go around a closed loop of some kind, that phase will either remain the same or ramp by some integer multiple of \(2\pi\).  The gradient of the phase is related to the velocity of the superfluid, so if the phase winds by \(2\pi\), that implies there is a circulation of flow and orbital angular momentum that has to be an integer multiple of \(\hbar\).  In the BKT theory, the demise of the superfluid phase as the system is warmed happens through the creation and unbinding of vortex-antivortex pairs.

On the other hand, the other recipients of the Buckley Prize were Gwendal Fève and Mike Manfra for their work (experiments here and here) regarding the braiding statistics of anyons in fractional quantum Hall systems.  I'd written about anyons here.  For electrons in 2D, the wavefunctions of excitations of the fractional quantum Hall system look like vortices.  The phase of the electronic wavefunction can wind due to circulation, and because electrons are charged, the phase can also wind due to magnetic flux attached to the little whirlpool.  It's the combination of these phase effects that can lead to those excitations acting like anyons (so that when two are physically swapped or braided around one another, the wavefunction picks up a phase factor that is not just the \(+1\) of bosons or the \(-1\) of fermions).  

As my friend Dan Arovas pointed out, there was a hope back in the early 1980s that perhaps vortices in superfluid helium would also act like anyons and have fractional statistics.  However, this paper by Haldane and Wu disproved that possibility.  

Vortex shedding, from here.
Because of the relationship between quantum phase winding and actual flow of (density) currents, vortices show up in lots of places in hard condensed matter physics.  Classical vortices are also physically nontrivial objects - they're topological and often seem to have very counterintuitive properties and motions.  Heck, Lord Kelvin was so taken by this that he thought (pre-quantum) that maybe everything is really vortices of some kind.  

Perhaps it is fitting that I am posting this on the 85th anniversary of the Tacoma Narrows bridge collapse.  That classic civil engineering failure was caused by vortex shedding by the bridge coupling to its torsional resonance frequency.  Vortices can have big consequences!  


November 07, 2025

Matt von HippelExplain/Teach/Advocate

Scientists have different goals when they communicate, leading to different styles, or registers, of communication. If you don’t notice what register a scientist is using, you might think they’re saying something they’re not. And if you notice someone using the wrong register for a situation, they may not actually be a scientist.

Sometimes, a scientist is trying to explain an idea to the general public. The point of these explanations is to give you appreciation and intuition for the science, not to understand it in detail. This register makes heavy use of metaphors, and sometimes also slogans. It should almost never be taken literally, and a contradiction between two different scientist explanations usually just means they are using incompatible metaphors for the same concept. Sometimes, scientists who do this a lot will comment on other metaphors you might have heard, referencing other slogans to help explain what those explanations miss. They do this knowing that they do, in the end, agree on the actual science: they’re just trying to give you another metaphor, with a deeper intuition for a neglected part of the story.

Other times, scientists are trying to teach a student to be able to do something. Teaching can use metaphors or slogans as introductions, but quickly moves past them, because it wants to show the students something they can use: an equation, a diagram, a classification. If a scientist shows you any of these equations/diagrams/classifications without explaining what they mean, then you’re not the student they had in mind: they had designed their lesson for someone who already knew those things. Teaching may convey the kinds of appreciation and intuition that explanations for the general public do, but that goal gets much less emphasis. The main goal is for students with the appropriate background to learn to do something new.

Finally, sometimes scientists are trying to advocate for a scientific point. In this register, and only in this register, are they trying to convince people who don’t already trust them. This kind of communication can include metaphors and slogans as decoration, but the bulk will be filled with details, and those details should constitute evidence: they should be a structured argument, one that lays out, scientifically, why others should come to the same conclusion.

A piece that tries to address multiple audiences can move between registers in a clean way. But if the register jumps back and forth, or if the wrong register is being used for a task, that usually means trouble. That trouble can be simple boredom, like a scientist’s typical conference talk that can’t decide whether it just wants other scientists to appreciate the work, whether it wants to teach them enough to actually use it, or whether it needs to convince any skeptics. It can also be more sinister: a lot of crackpots write pieces that are ostensibly aimed at convincing other scientists, but are almost entirely metaphors and slogans, pieces good at tugging on the general public’s intuition without actually giving scientists anything meaningful to engage with.

If you’re writing, or speaking, know what register you need to use to do what you’re trying to do! And if you run into a piece that doesn’t make sense, consider that it might be in a different register than you thought.

November 06, 2025

Terence TaoMathematical exploration and discovery at scale

Bogdan Georgiev, Javier Gómez-Serrano, Adam Zsolt Wagner, and I have uploaded to the arXiv our paper “Mathematical exploration and discovery at scale“. This is a longer report on the experiments we did in collaboration with Google Deepmind with their AlphaEvolve tool, which is in the process of being made available for broader use. Some of our experiments were already reported on in a previous white paper, but the current paper provides more details, as well as a link to a repository with various relevant data such as the prompts used and the evolution of the tool outputs.

AlphaEvolve is a variant of more traditional optimization tools that are designed to extremize some given score function over a high-dimensional space of possible inputs. A traditional optimization algorithm might evolve one or more trial inputs over time by various methods, such as stochastic gradient descent, that are intended to locate increasingly good solutions while trying to avoid getting stuck at local extrema. By contrast, AlphaEvolve does not evolve the score function inputs directly, but uses an LLM to evolve computer code (often written in a standard language such as Python) which will in turn be run to generate the inputs that one tests the score function on. This reflects the belief that in many cases, the extremizing inputs will not simply be an arbitrary-looking string of numbers, but will often have some structure that can be efficiently described, or at least approximated, by a relatively short piece of code. The tool then works with a population of relatively successful such pieces of code, with the code from one generation of the population being modified and combined by the LLM based on their performance to produce the next generation. The stochastic nature of the LLM can actually work in one’s favor in such an evolutionary environment: many “hallucinations” will simply end up being pruned out of the pool of solutions being evolved due to poor performance, but a small number of such mutations can add enough diversity to the pool that one can break out of local extrema and discover new classes of viable solutions. The LLM can also accept user-supplied “hints” as part of the context of the prompt; in some cases, even just uploading PDFs of relevant literature has led to improved performance by the tool. Since the initial release of AlphaEvolve, similar tools have been developed by others, including OpenEvolve, ShinkaEvolve and DeepEvolve.

We tested this tool on a large number (67) of different mathematics problems (both solved and unsolved) in analysis, combinatorics, and geometry that we gathered from the literature, and reported our outcomes (both positive and negative) in this paper. In many cases, AlphaEvolve achieves similar results to what an expert user of a traditional optimization software tool might accomplish, for instance in finding more efficient schemes for packing geometric shapes, or locating better candidate functions for some calculus of variations problem, than what was previously known in the literature. But one advantage this tool seems to offer over such custom tools is that of scale, particularly when when studying variants of a problem that we had already tested this tool on, as many of the prompts and verification tools used for one problem could be adapted to also attack similar problems; several examples of this will be discussed below. The following graphic illustrates the performance of AlphaEvolve on this body of problems:

Another advantage of AlphaEvolve was robustness adaptability: it was relatively easy to set up AlphaEvolve to work on a broad array of problems, without extensive need to call on domain knowledge of the specific task in order to tune hyperparameters. In some cases, we found that making such hyperparameters part of the data that AlphaEvolve was prompted to output was better than trying to work out their value in advance, although a small amount of such initial theoretical analysis was helpful. For instance, in calculus of variation problems, one is often faced with the need to specify various discretization parameters in order to estimate a continuous integral, which cannot be computed exactly, by a discretized sum (such as a Riemann sum), which can be evaluated by computer to some desired precision. We found that simply asking AlphaEvolve to specify its own discretization parameters worked quite well (provided we designed the score function to be conservative with regards to the possible impact of the discretization error); see for instance this experiment in locating the best constant in functional inequalities such as the Hausdorff-Young inequality.

A third advantage of AlphaEvolve over traditional optimization methods was the interpretability of many of the solutions provided. For instance, in one of our experiments we sought to find an extremum to a functional inequality such as the Gagliardo–Nirenberg inequality (a variant of the Sobolev inequality). This is a relatively well-behaved optimization problem, and many standard methods can be deployed to obtain near-optimizers that are presented in some numerical format, such as a vector of values on some discretized mesh of the domain. However, when we applied AlphaEvolve to this problem, the tool was able to discover the exact solution (in this case, a Talenti function), and create code that sampled from that function on a discretized mesh to provide the required input for the scoring function we provided (which only accepted discretized inputs, due to the need to compute the score numerically). This code could be inspected by humans to gain more insight as to the nature of the optimizer. (Though in some cases, AlphaEvolve’s code would contain some brute force search, or a call to some existing optimization subroutine in one of the libraries it was given access to, instead of any more elegant description of its output.)

For problems that were sufficiently well-known to be in the training data of the LLM, the LLM component of AlphaEvolve often came up almost immediately with optimal (or near-optimal) solutions. For instance, for variational problems where the gaussian was known to be the extremizer, AlphaEvolve would frequently guess a gaussian candidate during one of the early evolutions, and we would have to obfuscate the problem significantly to try to conceal the connection to the literature in order for AlphaEvolve to experiment with other candidates. AlphaEvolve would also propose similar guesses for other problems for which the extremizer was not known. For instance, we tested this tool on the sum-difference exponents of relevance to the arithmetic Kakeya conjecture, which can be formulated as a variational entropy inequality concerning certain two-dimensional discrete random variables. AlphaEvolve initially proposed some candidates for such variables based on discrete gaussians, which actually worked rather well even if they were not the exact extremizer, and already generated some slight improvements to previous lower bounds on such exponents in the literature. Inspired by this, I was later able to rigorously obtain some theoretical results on the asymptotic behavior on such exponents in the regime where the number of slopes was fixed, but the “rational complexity” of the slopes went to infinity; this will be reported on in a separate paper.

Perhaps unsurprisingly, AlphaEvolve was extremely good at locating “exploits” in the verification code we provided, for instance using degenerate solutions or overly forgiving scoring of approximate solutions to come up with proposed inputs that technically achieved a high score under our provided code, but were not in the spirit of the actual problem. For instance, when we asked it (link under construction) to find configurations to extremal geometry problems such as locating polygons with each vertex having four equidistant other vertices, we initially coded the verifier to accept distances that were equal only up to some high numerical precision, at which point AlphaEvolve promptly placed many of the points in virtually the same location so that the distances they determined were indistinguishable. Because of this, a non-trivial amount of human effort needs to go into designing a non-exploitable verifier, for instance by working with exact arithmetic (or interval arithmetic) instead of floating point arithmetic, and taking conservative worst-case bounds in the presence of uncertanties in measurement to determine the score. For instance, in testing AlphaEvolve against the “moving sofa” problem and its variants, we designed a conservative scoring function that only counted those portions of the sofa that we could definitively prove to stay inside the corridor at all times (not merely the discrete set of times provided by AlphaEvolve to describe the sofa trajectory) to prevent it from exploiting “clipping” type artefacts. Once we did so, it performed quite well, for instance rediscovering the optimal “Gerver sofa” for the original sofa problem, and also discovering new sofa designs for other problem variants, such as a 3D sofa problem.

For well-known open conjectures (e.g., Sidorenko’s conjecture, Sendov’s conjecture, Crouzeix’s conjecture, the ovals problem, etc.), AlphaEvolve generally was able to locate the previously known candidates for optimizers (that are conjectured to be optimal), but did not locate any stronger counterexamples: thus, we did not disprove any major open conjecture. Of course, one obvious possible explanation for this is that these conjectures are in fact true; outside of a few situations where there is a matching “dual” optimization problem, AlphaEvolve can only provide one-sided bounds on such problems and so cannot definitively determine if the conjectural optimizers are in fact the true optimizers. Another potential explanation is that AlphaEvolve essentially tried all the “obvious” constructions that previous researchers working on these problems had also privately experimented with, but did not report due to the negative findings. However, I think there is at least value in using these tools to systematically record negative results (roughly speaking, that a search for “obvious” counterexamples to a conjecture did not disprove the claim), which currently only exist as “folklore” results at best. This seems analogous to the role LLM Deep Research tools could play by systematically recording the results (both positive and negative) of automated literature searches, as a supplement to human literature review which usually reports positive results only. Furthermore, when we shifted attention to less well studied variants of famous conjectures, we were able to find some modest new observations. For instance, while AlphaEvolve only found the standard conjectural extremizer {z^n-1} to Sendov’s conjecture, as well as for variants such as Borcea’s conjecture, Schmeisser’s conjecture, or Smale’s conjecture it did reveal some potential two-parameter extensions to a conjecture of de Bruin and Sharma that had not previously been stated in the literature. (For this problem, we were not directly optimizing some variational scalar quantity, but rather a two-dimensional range of possible values, which we could adapt the AlphaEvolve framework to treat). In the future, I can imagine such tools being a useful “sanity check” when proposing any new conjecture, in that it will become common practice to run one of these tools against such a conjecture to make sure there are no “obvious” counterexamples (while keeping in mind that this is still far from conclusive evidence in favor of such a conjecture).

AlphaEvolve did not perform equally well across different areas of mathematics. When testing the tool on analytic number theory problems, such as that of designing sieve weights for elementary approximations to the prime number theorem, it struggled to take advantage of the number theoretic structure in the problem, even when given suitable expert hints (although such hints have proven useful for other problems). This could potentially be a prompting issue on our end, or perhaps the landscape of number-theoretic optimization problems is less amenable to this sort of LLM-based evolutionary approach. On the other hand, AlphaEvolve does seem to do well when the constructions have some algebraic structure, such as with the finite field Kakeya and Nikodym set problems, which we will turn to shortly.

For many of our experiments we worked with fixed-dimensional problems, such as trying to optimally pack {n} shapes in a larger shape for a fixed value of {n}. However, we found in some cases that if we asked AlphaEvolve to give code that took parameters such as {n} as input, and tested the output of that code for a suitably sampled set of values of {n} of various sizes, then it could sometimes generalize the constructions it found for small values of this parameter to larger ones; for instance, in the infamous sixth problem of this year’s IMO, it could use this technique to discover the optimal arrangement of tiles, which none of the frontier models could do at the time (although AlphaEvolve has no capability to demonstrate that this arrangement was, in fact, optimal). Another productive use case of this technique was for finding finite field Kakeya and Nikodym sets of small size in low-dimensional vector spaces over finite fields of various sizes. For Kakeya sets in {{\mathbf F}_q^d}, it located the known optimal construction based on quadratic residues in two dimensions, and very slightly beat (by an error term of size {O(q)}) the best construction in three dimensions; this was an algebraic construction (still involving quadratic residues) discovered empirically that we could then prove to be correct by first using Gemini’s “Deep Think” tool to locate an informal proof, which we could then convert into a formalized Lean proof by using Google Deepmind’s “AlphaProof” tool. At one point we thought it had found a construction in four dimensions which achieved a more noticeable improvement (of order {O(q^3)}) of what we thought was the best known construction, but we subsequently discovered that essentially the same construction had appeared already in a paper of Bukh and Chao, although it still led to a more precise calculation of the error term (to accuracy {O(q^{3/2})} rather than {O(q^2)}, where the error term now involves the Lang-Weil inequality and is unlikely to have a closed form). Perhaps AlphaEvolve had somehow absorbed the Bukh-Chao construction within its training data to accomplish this. However, when we tested the tool on Nikodym sets (which are expected to have asymptotic density {1}, although this remains unproven), it did find some genuinely new constructions of such sets in three dimensions, based on removing quadratic varieties from the entire space. After using “Deep Think” again to analyze these constructions, we found that they were inferior to a purely random construction (which in retrospect was an obvious thing to try); however, they did inspire a hybrid construction in which one removed random quadratic varieties and performed some additional cleanup, which ends up outperforming both the purely algebraic and purely random constructions. This result (with completely human-generated proofs) will appear in a subsequent paper.

Scott Aaronson On keeping a packed suitcase

Update (Nov. 6): I’ve closed the comments, as they crossed the threshold from “sometimes worthwhile” to “purely abusive.” As for Mamdani’s victory: as I like to say in such cases (and said, e.g., after George W. Bush’s and Trump’s victories), the silver lining to which I cling is that either I’ll be pleasantly surprised, and things won’t be quite as terrible as I expect, or else I’ll be vindicated.


This Halloween, I didn’t need anything special to frighten me. I walked all day around in a haze of fear and depression, unable to concentrate on my research or anything else. I saw people smiling, dressed up in costumes, and I thought: how?

The president of the Heritage Foundation, the most important right-wing think tank in the United States, has now explicitly aligned himself with Tucker Carlson, even as the latter has become a full-on Holocaust-denying Hitler-loving antisemite, who nods in agreement with the openly neo-Nazi Nick Fuentes. Meanwhile, Vice President J.D. Vance—i.e., plausibly the next President of the United States—pointedly did nothing whatsoever to distance himself from the MAGA movement’s lunatic antisemites, in response to their lunatic antisemitic questions at the Turning Point USA conference. (Vance thus dishonored the memory of Charlie Kirk, who for all my many disagreements with him, was a firmly committed Zionist.) It’s become undeniable that, once Trump himself leaves the stage, this is the future of MAGA, and hence of the Republican Party itself. Exactly as I warned would happen a decade ago, this is what’s crawled out from underneath the rock that Trump gleefully overturned.

While the Republican Party is being swallowed by a movement that holds that Jews like me have no place in America, the Democratic Party is being swallowed by a movement that holds that Jews have no place in Israel. If these two movements ever merged, the obvious “compromise” would be the belief, popular throughout history, that Jews have no place anywhere on earth.

Barring a miracle, New York City—home to the world’s second-largest Jewish community—is about to be led by a man for whom eradicating the Jewish state is his deepest, most fundamental moral imperative, besides of course the proletariat seizing the means of production. And to their eternal shame, something like 29% of New York’s Jews are actually going to vote for this man, believing that their own collaboration with evil will somehow protect them personally—in breathtaking ignorance of the millennia of Jewish history testifying to the opposite.

Despite what you might think, I try really, really hard not to hyperventilate or overreact. I know that, even if I lived in literal Warsaw in 1939, it would still be incumbent on me to assess the situation calmly and figure out the best response.

So for whatever it’s worth: no, I don’t expect that American Jews, even pro-Zionist Jews in New York City, will need to flee their homes just yet. But it does seem to me that they (to say nothing of British and Canadian and French Jews) might, so to speak, want to keep their suitcases packed by the door, as Jews have through the centuries in analogous situations. As Tevye says near the end of Fiddler on the Roof, when the Jews are given three days to evacuate Anatevka: “maybe this is why we always keep our hats on.” Diaspora Jews like me might also want to brush up on Hebrew. We can thank Hashem or the Born Rule that, this time around, at least the State of Israel exists (despite the bloodthirsty wish of half the world that it cease to exist), and we can reflect that these contingencies are precisely why Israel was created.


Let me make something clear: I don’t focus so much on antisemitism only because of parochial concern for the survival of my own kids, although I freely admit to having as much such concern as the next person. Instead, I do so because I hold with David Deutsch that, in Western civilization, antisemitism has for millennia been the inevitable endpoint toward which every bad idea ultimately tends. It’s the universal bad idea. It’s bad-idea-complete. Antisemitism is the purest possible expression of the worldview of the pitchfork-wielding peasant, who blames shadowy elites for his own failures in life, and who dreams in his resentment and rage of reversing the moral and scientific progress of humanity by slaughtering all those responsible for it. Hatred of high-achieving Chinese and Indian immigrants, and of gifted programs and standardized testing, are other expressions of the same worldview.

As far as I know, in 3,000 years, there hasn’t been a single example—not one—of an antisemitic regime of which one could honestly say: “fine, but once you look past what they did to the Jews, they were great for everyone else!” Philosemitism is no guarantee of general goodness (as we see for example with Trump), but antisemitism pretty much does guarantee general awfulness. That’s because antisemitism is not merely a hatred, but an entire false theory of how the world works—not just a but the conspiracy theory—and as such, it necessarily prevents its believers from figuring out true explanations for society’s problems.


I’d better end a post like this on a note of optimism. Yes, every single time I check my phone, I’m assaulted with twenty fresh examples of once-respected people and institutions, all across the political spectrum, who’ve now fallen to the brain virus, and started blaming all the world’s problems on “bloodsucking globalists” or George Soros or Jeffrey Epstein or AIPAC or some other suspicious stand-in du jour. (The deepest cuts come from the new Jew-haters who I myself once knew, or admired, or had some friendly correspondence with.)

But also, every time I venture out into the real world, I meet twenty people of all backgrounds whose brains still seem perfectly healthy, and who respond to events in a normal human way. Even in the dark world behind the screen, I can find dozens of righteous condemnations of Zohran Mamdani and Tucker Carlson and the Heritage Foundation and the others who’ve chosen to play footsie with those seeking a new Final Solution to the Jewish Question. So I reflect that, for all the battering it’s taken in this age of TikTok and idiocracy—even then, our Enlightenment civilization still has a few antibodies that are able to put up a fight.

In their beautiful book Abundance, Ezra Klein and Derek Thompson set out an ambitious agenda by which the Democratic Party could reinvent itself and defeat MAGA, not by indulging conspiracy theories but by creating actual broad prosperity. Their agenda is full of items like: legalizing the construction of more housing where people actually want to live; repealing the laws that let random busybodies block the construction of mass transit; building out renewable energy and nuclear; investing in science and technology … basically, doing all the things that anyone with any ounce of economic literacy knows to be good. The abundance agenda isn’t only righteous and smart: for all I know, it might even turn out to be popular. It’s clearly worth a try.

Last week I was amused to see Kate Willett and Briahna Joy Gray, two of the loudest voices of the conspiratorial far left, denounce the abundance agenda as … wait for it … a cover for Zionism. As far as they’re concerned, the only reason why anyone would talk about affordable housing or high-speed rail is to distract the masses from the evil Zionists murdering Palestinian babies in order to harvest their organs.

The more I thought about this, the more I realized that Willett and Gray actually have a point. Yes, solving America’s problems with reason and hard work and creativity, like the abundance agenda says to do, is the diametric opposite of blaming all the problems on the perfidy of Jews or some other scapegoat. The two approaches really are the logical endpoints of two directly competing visions of reality.

Naturally I have a preference between those visions. So I’ve been on a bit of a spending spree lately, in support of sane, moderate, pro-abundance, anti-MAGA, liberal Enlightenment forces retaking America. I donated $1000 to Alex Bores, who’s running for Congress in NYC, and who besides being a moderate Democrat who favors all the usual good things, is also a leader in AI safety legislation. (For more, see this by Eric Neyman of Alignment Research Center, or this from Scott Alexander himself—the AI alignment community has been pretty wowed.) I also donated $1000 to Scott Wiener, who’s running for Nancy Pelosi’s seat in California, has a nuanced pro-two-states, anti-Netanyahu position that causes him to get heckled as a genocidal Zionist, and authored the excellent SB1047 AI safety bill, which Gavin Newsom unfortunately vetoed for short-term political reasons. And I donated $1000 to Vikki Goodwin, a sane Democrat who’s running to unseat Lieutenant Governor Dan Patrick in my own state of Texas. Any other American office-seeker who resonates with this post, and who’d like a donation, can feel free to contact me as well.

My bag is packed … but for now, only for a brief trip to give the physics colloquium at Harvard, after which I’ll return back home to Austin. Until it becomes impossible, I call on my thousands of thoughtful, empathetic American readers to stay right where you are, and simply do your best to fight the brain-eaten zombies of both left and right. If you are one of the zombies, of course, then my calling you one doesn’t even begin to express my contempt: may you be remembered by history alongside the willing dupes of Hitler, Stalin, and Mao. May the good guys prevail.

Oh, and speaking of zombies, Happy Halloween everyone! Boooooooo!

Scott Aaronson UT Austin’s Statement on Academic Integrity

A month ago William Inboden, the provost of UT Austin (where I work), invited me to join a university-wide “Faculty Working Group on Academic Integrity.” The name made me think that it would be about students cheating on exams and the like. I didn’t relish the prospect but I said sure.

Shortly afterward, Jim Davis, the president of UT Austin, sent out an email listing me among 21 faculty who had agreed to serve on an important working group to decide UT Austin’s position on academic free speech and the responsibilities of professors in the classroom (!). Immediately I started getting emails from my colleagues, thanking me for my “service” and sharing their thoughts about what this panel needed to say in response to the Trump administration’s Compact on Higher Education. For context: the Compact would involve universities agreeing to do all sorts of things that the Trump administration wants—capping international student enrollment, “institutional neutrality,” freezing tuition, etc. etc.—in exchange for preferential funding. UT Austin was one of nine universities originally invited to join the Compact, along with MIT, Penn, Brown, Dartmouth, and more, and is the only one that hasn’t yet rejected it. It hasn’t accepted it either.

Formally, it was explained to me, UT’s Working Group on Academic Integrity had nothing to do with Trump’s Compact, and no mandate to either accept or reject it. But it quickly became obvious to me that my faculty colleagues would see everything we did exclusively in light of the Compact, and of other efforts by the Trump administration and the State of Texas to impose conservative values on universities. While not addressing current events directly, what we could do would be to take a strong stand for academic freedom, and more generally, for the role of intellectually independent universities in a free society.

So, led by Provost Inboden, over two meetings and a bunch of emails we hashed out a document. You can now read the Texas Statement on Academic Integrity, and I’d encourage you to do so. The document takes a pretty strong swing for academic freedom:

Academic freedom lies at the core of the academic enterprise.  It is foundational to the excellence of the American higher education system, and is non-negotiable. In the words of the U.S. Supreme Court, academic freedom is “a special concern of the First Amendment.” The world’s finest universities are in free societies, and free societies honor academic freedom.

The statement also reaffirms UT Austin’s previous commitments to the Chicago Principles of Free Expression, and the 1940 and 1967 academic freedom statements of the American Association of University Professors.

Without revealing too much about my role in the deliberations, I’ll say that I was especially pleased by the inclusion of the word “non-negotiable.” I thought that that word might acquire particular importance, and this was confirmed by the headline in yesterday’s Chronicle of Higher Education: As Trump’s Compact Looms, UT-Austin Affirms ‘Non-Negotiable’ Commitment to Academic Freedom (warning: paywall).

At the same time, the document also talks about the responsibility of a public university to maintain the trust of society, and about the responsibilities of professors in the classroom:

Academic integrity obligates the instructor to protect every student’s academic freedom and right to learn in an environment of open inquiry. This includes the responsibilities:

  • to foster classroom cultures of trust in which all students feel free to voice their questions and beliefs, especially when those perspectives might conflict with those of the instructor or other students;
  • to fairly present differing views and scholarly evidence on reasonably disputed matters and unsettled issues;
  • to equip students to assess competing theories and claims, and to use reason and appropriate evidence to form their own conclusions about course material; and
  • to eschew topics and controversies that are not germane to the course.

All stuff that I’ve instinctively followed, in nearly 20 years of classroom teaching, without the need for any statement telling me to. Whatever opinions I might get goaded into expressing on this blog about Trump, feminism, or Israel/Palestine, I’ve always regarded the classroom as a sacred space. (I have hosted a few fierce classroom debates about the interpretation of quantum mechanics, but even there, I try not to tip my own hand!)

I’m sure that there are commenters, on both ends of the political spectrum, who will condemn me for my participation in the faculty working group, and for putting my name on the statement. At this point in this blog’s history, commenters on both ends of the political spectrum would condemn me for saying that freshly baked chocolate chip cookies are delicious. But I like the statement, and find nothing in it that any reasonable person should disagree with. Overall, my participation in this process increased my confidence that UT Austin will be able to navigate this contentious time for the state, country, and world while maintaining its fundamental values. It made me proud to be a professor here.

n-Category Café The Inverse Cube Force Law

Here’s a draft of my next column for the Notices of the American Mathematical Society. It’s about the inverse cube force law in classical mechanics.

Newton’s Principia is famous for his investigations of the inverse square force law for gravity. But in this book Newton also did something that was rarely discussed until the 1990s. He figured out what kind of central force exerted upon a particle can rescale its angular velocity by a constant factor without affecting its radial motion. This turns out to be a force obeying an inverse cube law.

Given a particle in Euclidean space, a central force is a force that points toward or away from the origin and depends only on the particle’s distance from the origin. If the particle’s position at time is r(t) n\mathbf{r}(t) \in \mathbb{R}^n and its mass is some number m>0,m \gt 0, we have

mr¨(t)=F(r(t))r^(t) m \, \ddot{\mathbf{r}}(t) = F(r(t)) \,\hat{\mathbf{r}}(t)

where r^(t)\hat{\mathbf{r}}(t) is a unit vector pointing outward from the origin at the point r(t).\mathbf{r}(t). A particle obeying this equation always moves in a plane through the origin, so we can use polar coordinates and write the particle’s position as (r(t),θ(t).(r(t), \theta(t). With some calculation one can show the particle’s distance from the origin, r(t),r(t), obeys

mr¨(t)=F(r(t))+L 2mr(t) 3(1) m \ddot r(t) = F(r(t)) + \frac{L^2}{m r(t)^3 } \qquad \qquad (1)

Here L=mr(t) 2θ˙(t)L = m r(t)^2 \dot \theta(t), the particle’s angular momentum, is constant in time. The second term in the equation above says that the particle’s distance from the origin changes as if there were an additional force pushing it outward. This is a “fictitious force”, an artifact of working in polar coordinates. It is called the centrifugal force. And it obeys an inverse cube force law!

This explains Newton’s observation. Let us see why. Suppose we have two particles moving in two different central forces F 1F_1 and F 2,F_2, each obeying a version of equation (1), with the same mass mm and the same radial motion r(t),r(t), but different angular momenta L 1L_1 and L 2.L_2. Then we must have

F 1(r(t))+L 1 2mr(t) 3=F 2(r(t))+L 2 2mr(t) 3 F_1(r(t)) + \frac{L_1^2}{m r (t)^3} = F_2(r(t)) + \frac{L_2^2}{m r(t)^3}

If the particle’s angular velocities are proportional we must have L 2=kL 1L_2 = k L_1 for some constant k,k, so

F 2(r 1(t))F 1(r(t))=(k1)L 1mr(t) 3 F_2(r_1(t)) - F_1(r(t)) = \frac{(k - 1)L_1}{m r (t)^3}

This says that F 2F_2 equals F 1F_1 plus an additional inverse cube force.

There are other interesting things about the inverse cube force law. Newtonian gravity is an attractive inverse square force, say F(r)=c/r 2F(r) = -c/r^2 with c>0,c \gt 0, so in this case we have

mr¨(t)=c/r(t) 2+L 2mr(t) 3 m \ddot r(t) = -c/r(t)^2 + \frac{L^2}{m r(t)^3 }

Because 1/r 31/r^3 grows faster than 1/r 21/r^2 as r0,r \downarrow 0, as long as the angular momentum LL is nonzero the repulsion of the centrifugal force will beat the attraction of gravity for sufficiently small r,r, and the particle will not fall in to the origin. The same is true for any attractive force F(r)=c/r pF(r) = -c/r^p with p<3.p \lt 3. But an attractive inverse cube force can overcome the centrifugal force and make a particle fall in to the origin.

In fact there are three qualitatively different possibilities for the motion of a particle in an attractive inverse cube force F(r)=c/r 3,F(r) = -c/r^3, depending on the value of cc. With work we can solve for 1/r1/r as a function of θ\theta (which is easier than solving for rr). There are three cases depending on the value of

ω 2=1cmL 2 \omega^2 = 1 - \frac{c m}{L^2}

They are vaguely analogous to the elliptical, parabolic and hyperbolic orbits of a particle in an inverse square force law:

1r(θ)={Acos(ωθ)+Bsin(ωθ) if ω 2>0 A+Bθ if ω=0 Ae ωθ+Be ωθ if ω 2<0 \frac{1}{r(\theta)} = \left\{ \begin{array}{lcl} A \cos(\omega \theta) + B \sin(\omega \theta) & \text{if} & \omega^2 \gt 0 \\ \\ A + B \theta & \text{if} & \omega = 0 \\ \\ A e^{\omega \theta} + B e^{-\omega \theta} & \text{if} & \omega^2 \lt 0 \end{array} \right.

The third case occurs when the attractive inverse cube force is strong enough to overcome the centrifugal force: c>L 2/m.c \gt L^2/m. Then the particle can spiral in to its doom, hitting the origin in a finite amount of time after infinitely many orbits, like this:

All three curves in the equation above are called Cotes spirals, after Roger Cotes’ work on the inverse cube force law, published posthumously in 1722. Cotes seems to have been the first to compute the derivative of the sine function. After Cotes’ death at the age of 33, Newton supposedly said “If he had lived we would have known something.”

The subtlety of the inverse cube force law is vastly heightened when we study it using quantum rather than classical mechanics. Here if cc is too large the theory is ill-defined, because there is no reasonable choice of self-adjoint Hamiltonian. If cc is smaller the theory is well-behaved. But at a certain borderline point it exhibits a remarkable property: spontaneous breaking of scaling symmetry. I hope to discuss this in my next column.

For more on the inverse cube force law, see:

  • N. Grossman, The Sheer Joy of Celestial Mechanics, Birkhäuser, Basel, 1996, p. 34.

For more on Newton’s work involving the inverse cube force law, see:

Cotes’ book is

  • Roger Cotes, Harmonia Mensuarum, Cambridge, 1722.

November 04, 2025

n-Category Café Dynamics in Jordan Algebras

In ordinary quantum mechanics, in the special case where observables are described as self-adjoint n×nn \times n complex matrices, we can describe time evolution of an observable O(t)O(t) using Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = -i [H, O(t)]

where HH is a fixed self-adjoint matrix called the Hamiltonian. This framework is great when we want to focus on observables rather than states. But Heisenberg’s equation doesn’t make sense in a general Jordan algebra. In this stripped-down framework, all we can do is raise observables to powers and take real linear combinations of them. This lets us define a ‘Jordan product’ of observables:

AB=12((A+B) 2A 2B 2)=12(AB+BA) A \circ B = \frac{1}{2} ((A + B)^2 - A^2 - B^2) = \frac{1}{2} (A B + B A)

but not commutators and not multiplication by ii. What do we do then?

I wrote a long paper about this:

My starting-point was that self-adjoint complex matrices form not only a Jordan algebra with product

AB=12(AB+BA) A \circ B = \frac{1}{2} (A B + B A)

but also a Lie algebra with bracket

i[A,B]=i(ABBA) -i [A, B] = -i(A B - B A)

See, the commutator of two self-adjoint matrices is skew-adjoint, but we can multiply it by ii, or more conventionally i-i, to get something self-adjoint. That’s what is going on in Heisenberg’s equation. But this trick doesn’t work for other Jordan algebras, at least not automatically—so there was a lot to say.

I just bumped into a nice paper on this issue that I hadn’t seen before:

The idea here is pretty wild: you can replace the commutator in Heisenberg’s equation by an associator:

(A,B,C)=(AB)CA(BC) (A, B, C) = (A \circ B) \circ C - A \circ (B \circ C)

This is well-defined whenever our observables are elements in a Jordan algebra. Jordan algebras are always commutative, but rarely associative!

Here’s the trick. Let 𝔥 n()\mathfrak{h}_n(\mathbb{C}) be the Jordan algebra of self-adjoint n×nn \times n complex matrices, and let’s start with Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = -i [H, O(t)]

where H𝔥 n()H \in \mathfrak{h}_n(\mathbb{C}). Suppose we can write

H=4i[X,Y] H = -4i [X, Y]

for some X,Y𝔥 n()X, Y \in \mathfrak{h}_n(\mathbb{C}). In this case we can use a really cool identity to express the commutator in Heisenberg’s equation in terms of an associator:

[[X,Y],A]=14(X,A,Y) [[X, Y], A] = -\frac{1}{4}(X, A, Y)

This holds in any associative algebra if you define [X,Y]=XYYX[X,Y] = X Y - Y X, XY=12(XY+YX)X \circ Y = \tfrac{1}{2} (X Y + Y X) and (X,A,Y)=(X, A, Y) = (XA)YX(AY) (X \circ A) \circ Y - X \circ (A \circ Y). It’s easy to check: just expand out both sides and compare them!

Using this identity, we get

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

Now we’re describing dynamics using only operations that are available in any Jordan algebra!

This raises the question of when a self-adjoint complex matrix HH can be written as 4i[X,Y]-4i [X, Y] for self-adjoint matrices X,YX, Y. This is true whenever HH is traceless, since 𝔰𝔲(n)\mathfrak{su}(n) is a compact simple real Lie algebra, and every element of such a Lie algebra is a commutator (as shown by Akhieser).

But any self-adjoint complex matrix HH is of the form H+λIH' + \lambda I where HH' is traceless, so writing H=4i[X,Y]H' = -4i[X,Y] we have

[H,O(t)]=[H+λI,O(t)]=[H,O(t)] [H, O(t)] = [H' + \lambda I, O(t)] = [H', O(t)] =4i[[X,Y],O(t)]=i(X,O(t),Y) = -4i [[X,Y], O(t)] = i (X, O(t), Y)

so we can rewrite Heisenberg’s equation as

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

Moreover, in any Jordan algebra, any pair of elements X,YX, Y determines a derivation (X,,Y)(X, \cdot, Y): see Section I.7 of Jacobson’s Structure and Representations of Jordan Algebras. In the finite-dimensional case there is no difficulty with exponentiating any derivation to obtain a one-parameter group of automorphisms. Thus, for any elements X,YX, Y of a finite-dimensional Jordan algebra, the solution of the above equation always determines a one-parameter group of Jordan algebra automorphisms! And this is just what we’d want for describing how observables change with time.

The are two obvious next questions: one mathematical and one more philosophical.

First, how many one-parameter groups of Jordan algebra automorphisms do we actually get out of solutions to

ddtO(t)=(X,O(t),Y) \frac{d}{d t} O(t) = (X, O(t), Y)

In the case of 𝔥 n()\mathfrak{h}_n(\mathbb{C}), we get them all, since it’s already known that we get them all from Heisenberg’s equation

ddtO(t)=i[H,O(t)] \frac{d}{d t} O(t) = - i [H , O(t)]

What about 𝔥 n()\mathfrak{h}_n(\mathbb{R}) and 𝔥 n()\mathfrak{h}_n(\mathbb{H})? I’m actually more interested in the exceptional Jordan algebra 𝔥 3(𝕆)\mathfrak{h}_3(\mathbb{O}), and here it seems we get them all! This was shown in a paper that’s fairly hard to find even though it’s available for free online:

It starts on page 214 of the PDF file.

(The editor of this journal has some crazy ideas, which has put off some people I’m talking to about this paper. But you can’t judge a paper by the journal it appeared in. Truini and Biedenharn are good — in fact Biedenharn is famous for helping discover an identity, the Biedenharn–Elliott identity, that amounts to the pentagon identity for the category of representations of SU(2)\text{SU}(2)! And the paper looks fine, as far as I can tell.)

Second, the more philosophical question: what does it mean to describe dynamics using not one observable, the Hamiltonian, but two? Perhaps the best way to tackle this is to try doing it, and seeing how it works. Note that this method is not just good for dynamics, but for any Lie group of symmetries.

November 03, 2025

n-Category Café Second Quantization and the Kepler Problem

The poet Blake wrote that you can see a world in a grain of sand. But even better, you can see a universe in an atom!

Bound states of hydrogen atom correspond to states of a massless quantum particle moving at the speed of light around the Einstein universe — a closed, static universe where space is a 3-sphere. We need to use a spin-½ particle to account for the spin of the electron. The states of the massless spin-½ particle where it forms a standing wave then correspond to the orbitals of the hydrogen atom. This explains the secret 4-dimensional rotation symmetry of the hydrogen atom.

In fact, you can develop this idea to the point of getting the periodic table of elements from a quantum field theory on the Einstein universe! I worked that out here:

but you can see a more gentle explanation in the following series of blog articles.

  • Part 1: a quick overview of Kepler’s work on atoms and the solar system, and more modern developments.

  • Part 2: why the eccentricity vector is conserved for a particle in an inverse square force, and what it means.

  • Part 3: why the momentum of a particle in an inverse square force moves around in a circle.

  • Part 4: why the 4d rotation group SO(4)\text{SO}(4) acts on bound states of a particle in an attractive inverse square force.

  • Part 5: quantizing the bound states of a particle in an attractive inverse square force, and getting the Hilbert space L 2(S 3)L^2(S^3) for bound states of a hydrogen atom, neglecting the electron’s spin.

  • Part 6: how the Duflo isomorphism explains quantum corrections to the hydrogen atom Hamiltonian.

  • Part 7: why the Hilbert space of bound states for a hydrogen atom including the electron’s spin is L 2(S 3) 2.L^2(S^3) \otimes \mathbb{C}^2.

  • Part 8: why L 2(S 3) 2L^2(S^3) \otimes \mathbb{C}^2 is also the Hilbert space for a massless spin-1/2 particle in the Einstein universe.

  • Part 9: a quaternionic description of the hydrogen atom’s bound states (a digression not needed for later parts).

  • Part 10: changing the complex structure on L 2(S 3) 2L^2(S^3) \otimes \mathbb{C}^2 to eliminate negative-energy states of the massless spin-1/2 particle, as often done.

  • Part 11: second quantizing the massless spin-1/2 particle and getting a quantum field theory on the Einstein universe, or alternatively a theory of collections of electrons orbiting a nucleus.

  • Part 12: obtaining the periodic table of elements from a quantum field theory on the Einstein universe.

I explain how the Hamiltonian of this quantum field theory has to be tweaked a bit to give the ‘Madelung rules’ for the order in which electrons get added as we go along the periodic table:

November 02, 2025

Doug NatelsonInteresting preprints: chirality-induced spin selectivity + quantum gravity

This continues to be a very busy time, but I wanted to point out two preprints that caught my eye this week.  Their subjects are completely disparate, but they stand out as essentially reviews written in a much more conversational tone than the usual literature.

The first is this preprint about chirality-induced spin selectivity, a subject that I've mentioned before on this blog.  There is now an extensive body of evidence (of varying quality) that there is a connection between structural chirality of molecules and their interactions with the spin angular momentum of electrons.  This includes monolayers of chiral molecules leading to net spin polarization of photoemitted electrons (here), a lot of electronic transport experiments involving chiral molecules and magnetic electrodes that seem to show spin-dependent transmission that is absent with achiral molecules, and even a chirality dependence of molecular adsorption kinetics on magnetic surfaces (here).  The preprint is a provocative discussion of the topic and possible mechanisms, and the importance of precision in the description of the various phenomena.

On a completely different topic, this preprint is a fun discussion about quantum gravity (!) and how condensed matter ideas of "the vacuum" can lead to insights about how quantum mechanics and gravity might need to play together.  One fun bit early on is a discussion of something I like to point out to my undergrad stat mech students:  A single hydrogen atom in a very very large box will apparently (if the usual stat mech formalism of partition functions is valid) be spontaneously ionized, even when the box (which presumably functions as a reservoir at temperature \(T\)) and atom are at temperatures faaaaaar below the energy scale for ionization.  This is discussed nicely in this 1966 article in the Journal of Chemical Education.  Anyway, I thought this was an interesting discussion from three condensed matter theorists.

November 01, 2025

Tommaso DorigoNovember First

Today is November 1st, the day dedicated to the dead, and I am in northern Sweden where daylight is scarce this time of the year. The two things conjure to arise thoughts of a darkish nature. 


[Above, a lousy picture taken this evening in a cemetery in Gammelstad, close to Lulea, in Norrbotten, Sweden. Sorry for the bad quality... Yet the landscape with all those small lights was really inspiring.]

read more

October 31, 2025

Matt von HippelFear of the Dark, Physics Version

Happy Halloween! I’ve got a yearly tradition on this blog of talking about the spooky side of physics. This year, we’ll think about what happens…when you turn off the lights.

Over history, astronomy has given us larger and larger views of the universe. We started out thinking the planets, Sun, and Moon were human-like, just a short distance away. Measuring distances, we started to understand the size of the Earth, then the Sun, then realized how much farther still the stars were from us. Gradually, we came to understand that some of the stars were much farther away than others. Thinkers like Immanuel Kant speculated that “nebulae” were clouds of stars like our own Milky Way, and in the early 20th century better distance measurements confirmed it, showing that Andromeda was not a nearby cloud, but an entirely different galaxy. By the 1960’s, scientists had observed the universe’s cosmic microwave background, seeing as far out as it was possible to see.

But what if we stopped halfway?

Since the 1920’s, we’ve known the universe is expanding. Since the 1990’s, we’ve thought that that expansion is speeding up: faraway galaxies are getting farther and farther away from us. Space itself is expanding, carrying the galaxies apart…faster than light.

That ever-increasing speed has a consequence. It means that, eventually, each galaxy will fly beyond our view. One by one, the other galaxies will disappear, so far away that light will not have had enough time to reach us.

From our perspective, it will be as if the lights, one by one, started to turn out. Each faraway light, each cloudy blur that hides a whirl of worlds, will wink out. The sky will get darker and darker, until to an astronomer from a distant future, the universe will appear a strangely limited place:

A single whirl of stars, in a deep, dark, void.

October 30, 2025

Doug NatelsonScience journalism - dark times

At this point it's old hat to decry the problems facing traditional news media.  Still, it is abundantly clear in our late stage capitalist society that there has been a collective business decision over the last 20+ years that, like local newspapers and television news, real science journalism is not a money maker.   Just a few examples:  Seventeen years ago, CNN cut its entire science, technology and environment reporting team.  In 2022, Popular Science ceased publication.  In 2023, National Geographic laid off their staff writers.  Last week, the Wall Street Journal laid off their science and health reporters.  

I have it on good authority that there is now only one science reporter left at the WSJ.  One, at a time when science and technology are more critically important to our rapidly changing society than ever, and there is enormous tumult in the US and elsewhere about how science is or is not supported and is or is not factoring into policy decisions.  All of this is happening at a time when public trust in science is falling.  (Check out this from Science Friday.)  

(updated for context) Leaving aside professional science outlets (the news sections of Science, Nature, and society publications like Physics Today, C&EN, Physics World, Chemistry World), there are some good publications out there, like Quanta and Nautilus (both founded by nonprofits). There are outstanding public writers of science, like Philip Ball, Helen Czerski, Katie Mack, Ethan Siegel, and many others (apologies for the incompleteness of this list).  There are some excellent freelance journalists.  The internet also means that there are many opportunities for great engagement.  For example, the videos from 3blue1brown are uniformly outstanding.  However, there are no filters, and the temptation to be click-baity or sensationalistic is problematic.  

I have no solutions to offer, except that I encourage you to support good science journalism and reporting when you see it.  It's important.

October 28, 2025

n-Category Café Applied Category Theory 2026

The next annual conference on applied category theory is in Estonia!

For more details, read on!

The conference particularly encourages participation from underrepresented groups. The organizers are committed to non-discrimination, equity, and inclusion. The code of conduct for the conference is available here.

Deadlines

  • Registration: TBA
  • Abstracts Due: 23 March 2026
  • Full Papers Due: 30 March 2026
  • Author Notification: 11 May 2026
  • Adjoint School: 29 June – 3 July 2026
  • Conference: 6 – 10 July 2026
  • Final versions of papers for proceedings due: TBA

Submissions

ACT2026 accepts submissions in English, in the following three tracks:

  1. Research

  2. Software demonstrations

  3. Teaching and communication

The detailed Call for Papers is available here.

Extended abstracts and conference papers should be prepared with LaTeX. For conference papers please use the EPTCS style files available here. The submission link is here.

Reviewing is single-blind, and we are not making public the reviews, reviewer names, the discussions nor the list of under-review submissions. This is the same as previous instances of ACT.

Program Committee Chairs

  • Geoffrey Cruttwell, Mount Allison University, Sackville
  • Priyaa Varshinee Srinivasan, Tallinn University of Technology, Estonia

Program Committee

  • Alexis Toumi, Planting Space
  • Bryce Clarke, Tallinn University of Technology
  • Barbara König, University of Duisburg-Essen
  • Bojana Femic, Serbian Academy of Sciences and Arts
  • Chris Heunen, The University of Edinburgh
  • Daniel Cicala, Southern Connecticut State University
  • Dusko Pavlovic, University of Hawaii
  • Evan Patterson, Topos Institute
  • Fosco Loregian, Tallinn University of Technology
  • Gabriele Lobbia, Università di Bologna
  • Georgios Bakirtzis, Institut Polytechnique de Paris
  • Jade Master, University of Strathclyde
  • James Fairbanks, University of Florida
  • Jonathan Gallagher, Hummingbird Biosciences
  • Joe Moeller, Caltech
  • Jules Hedges, University of Strathclyde
  • Julie Bergner, University of Virginia
  • Kohei Kishida, University of Illinois, Urbana-Champaign
  • Maria Manuel Clementino, CMUC, Universidade de Coimbra
  • Mario Román, University of Oxford
  • Marti Karvonen, University College London
  • Martina Rovelli, UMass Amherst
  • Masahito Hasegawa, Kyoto University
  • Matteo Capucci, University of Strathclyde
  • Michael Shulman, University of San Diego
  • Nick Gurski, Case Western Reserve University
  • Niels Voorneveld, Cybernetica
  • Paolo Perrone, University of Oxford
  • Peter Selinger, Dalhousie University
  • Paul Wilson, University of Southampton
  • Robin Cockett, University of Calgary
  • Robin Piedeleu, University College London
  • Rory Lucyshyn-Wright, Brandon University
  • Rose Kudzman-Blais, University of Ottawa
  • Ryan Wisnesky, Conexus AI
  • Sam Staton, University of Oxford
  • Shin-Ya Katsumata, Kyoto Sangyo University
  • Simon Willerton, University of Sheffield
  • Spencer Breiner, National Institute of Standards and Technology
  • Tai Danae Bradley, SandboxAQ
  • Titouan Carette, École Polytechnique
  • Tom Leinster, The University of Edinburgh
  • Walter Tholen, York University

Teaching & Communication

  • Selma Dündar-Coecke, University College London, Institute of Education
  • Ted Theodosopoulos, Nueva School

Organizing Committee

  • Pawel Sobocinski, Tallinn University of Technology
  • Priyaa Varshinee Srinivasan, Tallinn University of Technology
  • Sofiya Taskova, Tallinn University of Technology
  • Kristi Ainen, Tallinn University of Technology

Steering Committee

  • John Baez, University of California, Riverside
  • Bob Coecke, University of Oxford
  • Dorette Pronk, Dalhousie University
  • David Spivak, Topos Institute
  • Michael Johnson, Macquarie University
  • Simona Paoli, University of Aberdeen

October 27, 2025

John PreskillThe sequel

This October, fantasy readers are devouring a sequel: the final installment in Philip Pullman’s trilogy The Book of Dust. The series follows student Lyra Silvertongue as she journeys from Oxford to the far east. Her story features alternate worlds, souls that materialize as talking animals, and a whiff of steampunk. We first met Lyra in the His Dark Materials trilogy, which Pullman began publishing in 1995. So some readers have been awaiting the final Book of Dust volume for 30 years. 

Another sequel debuts this fall. It won’t spur tens of thousands of sales; nor will Michael Sheen narrate an audiobook version of it. Nevertheless, the sequel should provoke as much thought as Pullman’s: the sequel to the Maryland Quantum-Thermodynamics Hub’s first three years.

More deserving of a Carnegie Medal than our hub, but the hub deserves no less enthusiasm!

The Maryland Quantum-Thermodynamics Hub debuted in 2022, courtesy of a grant from the John F. Templeton Foundation. Six theorists, three based in Maryland, have formed the hub’s core. Our mission has included three prongs: research, community building, and outreach. During the preceding decade, quantum thermodynamics had exploded, but mostly outside North America. We aimed to provide a lodestone for the continent’s quantum-thermodynamics researchers and visitors.

Also, we aimed to identify the thermodynamics of how everyday, classical physics emerges from quantum physics. Quantum physics is reversible (doesn’t distinguish the past from the future), is delicate (measuring a quantum system can disturb it), and features counterintuitive phenomena such as entanglement. In contrast, our everyday experiences include irreversibility (time has an arrow), objectivity (if you and I read this article, we should agree about its contents), and no entanglement. How does quantum physics give rise to classical physics at large energy and length scales? Thermodynamics has traditionally described macroscopic, emergent properties. So quantum thermodynamics should inform our understanding of classical reality’s emergence from quantum mechanics.

Our team has approached this opportunity from three perspectives. One perspective centers on quantum Darwinism, a framework for quantifying how interactions spread information about an observed quantum system. Another perspective highlights decoherence, the contamination of a quantum system by its environment. The third perspective features incompatible exchanged quantities, described in an earlier blog post. Or two. Or at least seven

Each perspective led us to discover a tension, or apparent contradiction, that needs resolving. One might complain that we failed to clinch a quantum-thermodynamic theory of the emergence of classical reality. But physicists adore apparent contradictions as publishers love splashing “New York Times bestseller” on their book covers. So we aim to resolve the tensions over the next three years.

Physicists savor paradoxes and their ilk.

I’ll illustrate the tensions with incompatible exchanged quantities, of course. Physicists often imagine a small system, such as a quantum computer, interacting with a large environment, such as the surrounding air and the table on which the quantum computer sits. The system and environment may exchange energy, particles, electric charge, etc. Typically, the small system thermalizes, or reaches a state mostly independent of its initial conditions. For example, after exchanging enough energy with its environment, the system ends up at the environment’s temperature, mostly regardless of the system’s initial temperature. 

For decades, physicists implicitly assumed that the exchanged quantities are compatible: one can measure them simultaneously. But one can’t measure all of a quantum system’s properties simultaneously. Position and momentum form the most famous examples. Incompatibility epitomizes quantum physics, underlying Heisenberg’s uncertainty relation, quantum error correction, and more. So collaborators and I ask how exchanged quantities’ incompatibility alters thermalization, which helps account for time’s arrow. 

Our community has discovered that such incompatibility can hinder certain facets of thermalization—in a sense, stave off certain aspects of certain quantum systems’ experience of time. But incompatible exchanged quantities enhance other features of thermalization. How shall we reconcile the hindrances with the enhancements? Does one of the two effects win out? I hope to report back in three years. For now, I’m rooting for Team Hindrance.

In addition to resolving apparent conflicts, we’re adding a fourth perspective to our quiver—a gravitational one. In our everyday experiences, space-time appears smooth; unlike Lyra’s companion Will in The Subtle Knife, we don’t find windows onto other worlds. But quantum physics, combined with general relativity, suggests that you’d find spikes and dips upon probing space-time over extremely short length scales. How does smooth space-time emerge from its quantum underpinnings? Again, quantum thermodynamics should help us understand.

To address these challenges, we’re expanding the hub’s cast of characters. The initial cast included six theorists. Two more are joining the crew, together with the hub’s first two experimentalists. So is our first creative-writing instructor, who works at the University of Maryland (UMD) Jiménez-Porter Writers’ House.

As the hub has grown, so has the continent’s quantum-thermodynamics community. We aim to continue expanding that community and strengthening its ties to counterparts abroad. As Lyra learned in Pullman’s previous novel, partnering with Welsh miners and Czech book sellers and Smyrnan princesses can further one’s quest. I don’t expect the Maryland Quantum-Thermodynamics Hub to attract Smyrnan princesses, but a girl can dream. The hub is already partnering with the John F. Templeton Foundation, Normal Computing, the Fidelity Center for Applied Technology, the National Quantum Laboratory, Maryland’s Capital of Quantum team, and more. We aim to integrate quantum thermodynamics into North America’s scientific infrastructure, so that the field thrives here even after our new grant terminates. Reach out if you’d like to partner with us.

To unite our community, the hub will host a gathering—a symposium or conference—each year. One conference will feature quantum thermodynamics and quantum-steampunk creative writing. Scientists and authors will present. We hope that both groups will inspire each other, as physicist David Deutsch’s work on the many-worlds formulation of quantum theory inspired Pullman.

That conference will follow a quantum-steampunk creative-writing course to take place at UMD during spring 2026. I’ll co-teach the course with creative-writing instructor Edward Daschle. Students will study quantum thermodynamics, read published science-fiction stories, write quantum-steampunk stories, and critique each other’s writing. Five departments have cross-listed the course: physics, arts and humanities, computer science, chemistry, and mechanical engineering. If you’re a UMD student, you can sign up in a few weeks. Do so early; seats are limited! We welcome graduate students and undergrads, the latter of whom can earn a GSSP general-education credit.1 Through the course, the hub will spread quantum thermodynamics into Pullman’s world—into literature.

Pullman has entitled his latest novel The Rose Field. The final word refers to an object studied by physicists. A field, such as an electric or gravitational field, is a physical influence spread across space. Hence fiction is mirroring physics—and physics can take its cue from literature. As ardently as Lyra pursues the mysterious particle called Dust, the Maryland Quantum-Thermodynamics Hub is pursuing a thermodynamic understanding of the classical world’s emergence from quantum physics. And I think our mission sounds as enthralling as Lyra’s. So keep an eye on the hub for physics, community activities, and stories. The telling of Lyra’s tale may end this month, but the telling of the hub’s doesn’t.

1Just don’t ask me what GSSP stands for.

October 24, 2025

Matt von HippelC. N. Yang, Dead at 103

I don’t usually do obituaries here, but sometimes I have something worth saying.

Chen Ning Yang, a towering figure in particle physics, died last week.

Picture from 1957, when he received his Nobel

I never met him. By the time I started my PhD at Stony Brook, Yang was long-retired, and hadn’t visited the Yang Institute for Theoretical Physics in quite some time.

(Though there was still an office door, tucked behind the institute’s admin staff, that bore his name.)

The Nobel Prize doesn’t always honor the most important theoretical physicists. In order to get a Nobel Prize, you need to discover something that gets confirmed by experiment. Generally, it has to be a very crisp, clear statement about reality. New calculation methods and broader new understandings are on shakier ground, and theorists who propose them tend to be left out, or at best combined together into lists of partial prizes long after the fact.

Yang was lucky. With T. D. Lee, he had made that crisp, clear statement. He claimed that the laws of physics, counter to everyone’s expectations, are not the same when reflected in a mirror. In 1956, Wu confirmed the prediction, and Lee and Yang got the prize the year after.

That’s a huge, fundamental discovery about the natural world. But as a theorist, I don’t think that was Yang’s greatest accomplishment.

Yang contributed to other fields. Practicing theorists have seen his name strewn across concepts, formalisms, and theorems. I didn’t have space to talk about him in my article on integrability for Quanta Magazine, but only just barely: another paragraph or two, and he would have been there.

But his most influential contribution is something even more fundamental. And long-time readers of this blog should already know what it is.

Yang, along with Robert Mills, proposed Yang-Mills Theory.

There isn’t a Nobel prize for Yang-Mills theory. In 1953, when Yang and Mills proposed the theory, it was obviously wrong, a theory that couldn’t explain anything in the natural world, mercilessly mocked by famous bullshit opponent Wolfgang Pauli. Not even an ambitious idea that seemed outlandish (like plate tectonics), it was a theory with such an obvious missing piece that, for someone who prioritized experiment like the Nobel committee does, it seemed pointless to consider.

All it had going for it was that it was a clear generalization, an obvious next step. If there are forces like electromagnetism, with one type of charge going from plus to minus, why not a theory with multiple, interacting types of charge?

Nothing about Yang-Mills theory was impossible, or contradictory. Mathematically, it was fine. It obeyed all the rules of quantum mechanics. It simply didn’t appear to match anything in the real world.

But, as theorists learn, nature doesn’t let a good idea go to waste.

Of the four fundamental forces of nature, as it would happen, half are Yang-Mills theories. Gravity is different, electromagnetism is simpler, and could be understood without Yang and Mills’ insights. But the weak nuclear force, that’s a Yang-Mills theory. It wasn’t obvious in 1953 because it wasn’t clear how the massless, photon-like particles in Yang-Mills theory could have mass, and it wouldn’t become clear until the work of Peter Higgs over a decade later. And the strong nuclear force, that’s also a Yang-Mills theory, missed because of the ability of such a strong force to “confine” charges, hiding them away.

So Yang got a Nobel, not for understanding half of nature’s forces before anyone else had, but from a quirky question of symmetry.

In practice, Yang was known for all of this, and more. He was enormously influential. I’ve heard it claimed that he personally kept China from investing in a new particle collider, the strength of his reputation the most powerful force on that side of the debate, as he argued that a developing country like China should be investing in science with more short-term industrial impact, like condensed matter and atomic physics. I wonder if the debate will shift with his death, and what commitments the next Chinese five-year plan will make.

Ultimately, Yang is an example of what a theorist can be, a mix of solid work, counterintuitive realizations, and the thought-through generalizations that nature always seems to make use of in the end. If you’re not clear on what a theoretical physicist is, or what one can do, let Yang’s story be your guide.

Tommaso DorigoAre We Stochastic Parrots, Too? What LLMs Teach Us About Intelligence And Understanding

Having interacted for a few months with ChatGPT 5 now, both for work-related problems and for private / self-learning tasks, I feel I might share some thoughts here on what these large models can tell us about our own thought processes. 

The sentence above is basically giving away my bottomline from square one, but I suppose I can elaborate a bit more on the concept. LLMs have revolutionized a wide range of information-processing tasks in just three or four years. Looking back, the only comparable breakthrough I can recall is the advent of internet search engines in the early 1990s. But as exciting and awesome this breakthrough is, it inspires me still more to ponder on how this is even possible. Let me unpack this.

read more

October 15, 2025

Clifford JohnsonNobel Prize in Physics 2025: Who/What/Why

I started a tradition a little while back where every year we have a special departmental colloquium entitled "The Nobel Prize in Physics: Who/What/Why". This year my job in finding speakers was made easier by having 2/3 of this years newly-minted Nobel Prize winners in physics in the Department! (Michel Devoret and John Martinis.) So our room was a bit more well-attended than normal...(hundreds and hundreds rather than dozens and dozens). Here is a recording of the event, which I was delighted to host, and there's a celebration afterwards too. (Pls share widely!)
[...] Click to continue reading this post

The post Nobel Prize in Physics 2025: Who/What/Why appeared first on Asymptotia.

October 11, 2025

Doug NatelsonACS National Nanotechnology Day webinar, Thursday Oct 9

Time for a rare bit of explicit self-promotion on this blog.  This coming Thursday, October 9, as part of the American Chemical Society's activities for National Nanotechnology Day (Why October 9?  In US convention, Oct 9 = 10/9, and 10-9 m = 1 nm.  Look, it wasn't my idea....), I'm speaking in a free webinar titled "Illuminating the Nano Frontier", with Prof. Dongling Ma of INRS in Quebec.  The event is 11am-12:30pm EDT, and there will also be a recording for people who are unable to watch it live.  Should be a fun event. 

UpdateHere is the link to the webinar recording.  It's free and open-access.

October 08, 2025

Tommaso DorigoInterna

With this post I would like to present a short update of my personal life to the few readers who are interested in that topic. You know, when I started writing online (over 20 years ago!), blogs used to contain a much more personal, sometimes introspective, description of the owner's private life and actions. Since long, they have been substituted by much more agile, quick-to-consume videos. But the old-fashioned bloggers who stuck with that medium continue to have a life - albeit certainly a less glamorous one than that of today's influencers; so some reporting of personal affairs is not out of place here. 

read more

October 01, 2025

Robert HellingHolosplit

 Recently I had to update Mathematica on my laptop and after having solved the challenges of the license manager that keeps looking different every time I have to use it, I learned that Mathematica 14 can now officially work with finite fields.

This reminded me that for a while I wanted to revive an old project that had vanished together with the hard drive of some old computer: Holosplit. So, over the last two days and with the help of said version of Mathematica I did a complete rewrite which you can now find on Github.

It consists of two C programs "holosplit" and "holojoin". To the first you give a positive integer \(N\) and a file and it spits out a new file ("fragment") that is roughly \(1/N\) of the size. Every time you do that you obtain a new random fragment.

The later you give any collection of \(N\) of these fragments and it reproduces the original file. So you can for example distribute a file over 10 people such that when any 3 of them work together, they can recover the original. 

How does it work? I uses the finite field \(F\) of \(2^3=256\) elements (in the Github repository, there is also a header file that implements arithmetic in \(F\) and matrix operations like product and inverse over it). Each time, it is invoked, it picks a random vector \(v\in F^N\) and writes it to the output. Then it reads \(N\) bytes from the file at a time which it also interprets as a vector \(d\in F^N\). It then outputs the byte that corresponds to the scalar product \(v\cdot d\).

To reassemble the file, holojoin takes the \(N\) files with its random vectors \(v_1,\ldots,v_N\) and interprets those as the rows of a \(N\times N\) matrix \(A\). With probability

$$\frac{\prod_{k=1}^N \left(256^N-k\right)}{(256)^{N^2}}$$

which exponentially in \(N\) approaches 1 this matrix is invertible (homework: why?). So we can read one byte from each file, assemble those into yet another vector \(e\in F^N\) and recover

$$d=A^{-1}e.$$

Besides the mathematics, it also poses philosophical/legal questions: Consider for example the original file is copyrighted, for example an mp3 or a video. The fragments are clearly derived works. But individually, they do not contain the original work, without sufficiently many other fragments they are useless (although not in a cryptographic sense). So by publishing one fragment, I do not provide access to the original work. What if others publish other fragments? Then my fragment could be the last remaining one that was missing. If there are more, any individual fragment is redundant so publishing it strictly speaking does not provide new information. 

September 26, 2025

Peter Rohde Photo albums

Peter’s photos: https://www.icloud.com/sharedalbum/#B275oqs3qKSZvQ

Screenshots: https://www.icloud.com/sharedalbum/#B27532ODWjIQb9

Climbing book launch: https://www.icloud.com/sharedalbum/#B27GWZuqDGnuOyN

Salisbury waters: https://www.icloud.com/sharedalbum/#B275qXGF1JQFkx

Christmas with Ash: https://www.icloud.com/sharedalbum/#B27G6XBubAhoT6

Hosin BBQ duck: https://www.icloud.com/sharedalbum/#B27GY8gBYG3b5mD

Hawks Nest to Smiths Lake: https://www.icloud.com/sharedalbum/#B2759UlCqSH5bE

Europe & Alps: https://www.icloud.com/sharedalbum/#B275ON9t3W0lu

Point Perpendicular: https://www.icloud.com/sharedalbum/#B27GqkRUiGivXD2

Newnes canyoning: https://www.icloud.com/sharedalbum/#B27GfnH8tgHSmX

Coffs Harbour to Yamba: https://www.icloud.com/sharedalbum/#B27J0DiRHJKuuWr

Wendy Bruere Christmas (2020): https://www.icloud.com/sharedalbum/#B27G4TcsmGoHysj

Six Foot Track: https://www.icloud.com/sharedalbum/#B2753qWtHZA9EX

Kosciusko to Kiandra: https://www.icloud.com/sharedalbum/#B27GgZLKuGaewVm

Camping food: https://www.icloud.com/sharedalbum/#B27GtnIORgbmHu

The Aardvark: https://www.icloud.com/sharedalbum/#B275VaUrzvmAiT

Kangaroo Valley kayaking: https://www.icloud.com/sharedalbum/#B27JEsNWnJrCpi0

Claustral canyon: https://www.icloud.com/sharedalbum/#B2755Z2WMOTpsk

Budawang: https://www.icloud.com/sharedalbum/#B27GDdyTvGvpINL

Mother’s Day panoramas (2021): https://www.icloud.com/sharedalbum/#B27GFssfGG9WmJP

Point Perpendicular & Nowra: https://www.icloud.com/sharedalbum/#B27GRMtznGPdeuZ

Blood moon: https://www.icloud.com/sharedalbum/#B27GdIshaG8NgGX

La Perouse to Coogee: https://www.icloud.com/sharedalbum/#B275aVbMK4h7qo

Canberra ASPI launch: https://www.icloud.com/sharedalbum/#B27GQOeMmGj4Zcv

Edible foraging: https://www.icloud.com/sharedalbum/#B275ejO179Si0N

Sydney to Wollongong: https://www.icloud.com/sharedalbum/#B275M7GFPUasMe

Album for Dad, Father’s Day (2021): https://www.icloud.com/sharedalbum/#B2752plgjnnkUe

Vaucluse (with Cheryl, Nestor & Wendy): https://www.icloud.com/sharedalbum/#B275CmvAS4uA0Z

Bouddi National Park: https://www.icloud.com/sharedalbum/#B27GdPblXG8WdOo

Tom Thumb (the 2nd): https://www.icloud.com/sharedalbum/#B275aDWbr4CN2w

Eden to Victoria: https://www.icloud.com/sharedalbum/#B27GJDfWGArX8l

Wendy’s book launch (the 2nd): https://www.icloud.com/sharedalbum/#B27GIcgc2G7h08y

Mark & Pat Bruere visit Sydney: https://www.icloud.com/sharedalbum/#B27G0ehgLbyWyg

New Years Eve climb (2021): https://www.icloud.com/sharedalbum/#B27Ju8EH6JOZxmU

Newnes Canyoning (2022): https://www.icloud.com/sharedalbum/#B275BydzFU0GZ8

Royal National Park (2022): https://www.icloud.com/sharedalbum/#B27GlxzuqGVI5nE

Peter & Wendy: https://www.icloud.com/sharedalbum/#B27Gf693ZG52tfd

Book photo shoots: too rude…

Wendy & Peter’s mushroom trip: https://www.icloud.com/sharedalbum/#B27GrhkPxG27So8

Post-mushroom hike: https://www.icloud.com/sharedalbum/#B27GdFryYG8i3Ur

Wendy Kalymnos favourites: https://www.icloud.com/sharedalbum/#B27JqstnBJEXkH2

Wendy Frenchmans screenshots: https://www.icloud.com/sharedalbum/#B27Jr1PPdJpd7Dq

Instagram: https://www.icloud.com/sharedalbum/#B27GzFCC1Gb4tqr

Haute route: https://www.icloud.com/sharedalbum/#B27J8GySPJtWoQ1

Kim’s KKKalendar: https://www.icloud.com/sharedalbum/#B275fk75vIL0sH

Frenchmans Cap Wild: https://www.icloud.com/sharedalbum/#B27G4VTwGGoFBkz

Photoshoot with Zixin: https://www.icloud.com/sharedalbum/#B27GPCdxkGKPkM4

Wendy birthday hike (2023): https://www.icloud.com/sharedalbum/#B27GWBC59GnHpQW

Bateman’s Bay to Bawley Point: https://www.icloud.com/sharedalbum/#B27JsHvHoJ8bxWf

Stockton Sand dunes (2023): https://www.icloud.com/sharedalbum/#B27GVfZ2vGloFZV

Wendy book launch (2023): https://www.icloud.com/sharedalbum/#B27J058xyJR4IBM

Dolomites (2023): https://www.icloud.com/sharedalbum/#B0Z5kuVsbGJUzKO

Mount Arapiles: https://www.icloud.com/sharedalbum/#B275GH8Mq8Uh2X

Mount Solitary loop: https://www.icloud.com/sharedalbum/#B275nhQST2mETE

Klaus Hanz Franz Rohde Kunst: https://www.icloud.com/sharedalbum/#B27GqQrCLGiY3vb

Klaus Rohde funeral slideshow: https://www.icloud.com/sharedalbum/#B27GDZLe8GXP58K

Dad (old, B&W): https://www.icloud.com/sharedalbum/#B27GLLXGLJ5mbT2

Klaus & Ursula wedding: https://www.icloud.com/sharedalbum/#B275cLqfN7154g

Test Greece: https://www.icloud.com/sharedalbum/#B27Jq4WnLJ6JMNd

From Will Skea (Alps): https://www.icloud.com/sharedalbum/#B27JHciePJFwacG

From Will Skea (Frenchmans Cap): https://www.icloud.com/sharedalbum/#B275ZhN2v3EVq6

From Will Skea (Arapiles): https://www.icloud.com/sharedalbum/#B27JPrgBGJu3BTD

Coffs Harbour to Yamba (2): https://www.icloud.com/sharedalbum/#B27GFqhgJG9LHgT

Mark magic show (2021): https://www.icloud.com/sharedalbum/#B27G60dj6ARCvd

Wendy Christmas present (2020): https://www.icloud.com/sharedalbum/#B275FrPQ6GxvRu

AHS 25 year reunion: https://www.icloud.com/sharedalbum/#B275O3DjHUvSv

WhatsApp: https://www.icloud.com/sharedalbum/#B275tzEA5fX1nc

Armidale High School: https://www.icloud.com/sharedalbum/#B27GnbeumG4PnAF

Book photos for Mum & Dad: https://www.icloud.com/sharedalbum/#B27Gtec4XQkASe

Miscellaneous: https://www.icloud.com/sharedalbum/#B27Gq6kMgGKn7GR

Three Capes Trail (2022): https://www.icloud.com/sharedalbum/#B27G7HOIlGrDUGZ

Childhood computer programming: https://www.icloud.com/sharedalbum/#B275fu2MutDU8N

Magic with Mark in Maroubra: https://www.icloud.com/sharedalbum/#B27Gv6DhEGD9U3G

Photoshoot with Zixin (2024): https://www.icloud.com/sharedalbum/#B27GCATCnJGoRfW

Butt Crack (2021): https://www.icloud.com/sharedalbum/#B275VtHQfMv0zw

Greece photos new (edited to remove photos from wrong album): https://www.icloud.com/sharedalbum/#B27GY3uThGoBcGj

Singapore (all combined): https://www.icloud.com/sharedalbum/#B275qsTcwJKJjl

Hong Kong (transit): https://www.icloud.com/sharedalbum/#B2759v1AbS8Hve

Taiwan: https://www.icloud.com/sharedalbum/#B27GQD2D7Gw0hAp

India (combined): https://www.icloud.com/sharedalbum/#B27Gtue8VQy83g

Freycinet: https://www.icloud.com/sharedalbum/#B27G5VpecGE5Tbg

Triglav: https://www.icloud.com/sharedalbum/#B275MbK9Vy8erz

Shared with me: https://www.icloud.com/sharedalbum/#B27GGXqixzPOrm

Mount Wellington climbing: https://www.icloud.com/sharedalbum/#B27Gd59qiG8Kjy4

New Zealand combined (2004): https://www.icloud.com/sharedalbum/#B27GIZ8BIGNN5jy

New Zealand combined (2005): https://www.icloud.com/sharedalbum/#B27GcuRfIGFVIcL

Yea: https://www.icloud.com/sharedalbum/#B27GZYbYHGhFIir

Mount Pleasant: https://www.icloud.com/sharedalbum/#B275Iy2hC0JTTL

D’Aguilar: https://www.icloud.com/sharedalbum/#B27Gh7fzTGZBosS

Bali (2001): https://www.icloud.com/sharedalbum/#B27G1qNHBGOTbIr

Samba Ninjas: https://www.icloud.com/sharedalbum/#B27GG34bAzqQ0v

Armidale (misc): https://www.icloud.com/sharedalbum/#B27GSkLVwGyobbX

Emma’s party (2008): https://www.icloud.com/sharedalbum/#B275S2ms99Zyby

Goettingen (2011): https://www.icloud.com/sharedalbum/#B27JIrbT3Jsgxhd

South Coast track: https://www.icloud.com/sharedalbum/#B27G58NWBG6QyN7

Minsk (2006): https://www.icloud.com/sharedalbum/#B27G3JpSBGX1UkQ

Baden-Baden (2019): https://www.icloud.com/sharedalbum/#B27595X5HTVzJr

Berlin (combined): https://www.icloud.com/sharedalbum/#B27JqWzChJ6qizD

Switzerland (combined): https://www.icloud.com/sharedalbum/#B275zXwoYGJ6HMF

Italy highlights: https://www.icloud.com/sharedalbum/#B27G47PHQGoJium

Germany (misc): https://www.icloud.com/sharedalbum/#B275hPMfYGu5xVJ

Garmisch (2022): https://www.icloud.com/sharedalbum/#B27GFsbvlG9Xrr6

Germany (2019): https://www.icloud.com/sharedalbum/#B27G6Mn98G56Ncb

Garmisch (2006): https://www.icloud.com/sharedalbum/#B27J5lIdKGLC9KG

Baden-Baden (2005): https://www.icloud.com/sharedalbum/#B275sWRpHHQkt9

Berlin (2005): https://www.icloud.com/sharedalbum/#B27GgOQtrGjQrpH

Zugspitze (2005): https://www.icloud.com/sharedalbum/#B27G81mNdGcApGt

Amsterdam, Bristol (2006): https://www.icloud.com/sharedalbum/#B275B9SRzyBjlH

Baden-Baden (2006): https://www.icloud.com/sharedalbum/#B275eD9V79I2XR

Berlin (2006): https://www.icloud.com/sharedalbum/#B275toRf1fH8MD

Berlin, Jena (2007): https://www.icloud.com/sharedalbum/#B27GTI3fvGVgNit

Erlangen (2006): https://www.icloud.com/sharedalbum/#B27JrotZ2JpMb0i

Garmisch (2010): https://www.icloud.com/sharedalbum/#B27JPJPSiJurzNg

Germany (2010): https://www.icloud.com/sharedalbum/#B275FhYPQP650

Stuttgart (2006): https://www.icloud.com/sharedalbum/#B27GmitydGVVaZh

Changi (2019): https://www.icloud.com/sharedalbum/#B27GnmlKoG4JHpX

Japan (2007): https://www.icloud.com/sharedalbum/#B275AerZbG6FxVL

Japan (2012): https://www.icloud.com/sharedalbum/#B27GjBjobGg6PUa

Miscellaneous (including Japan 2013): https://www.icloud.com/sharedalbum/#B27GTpbybGySbE8

Currumbin & Tugin (2021): https://www.icloud.com/sharedalbum/#B275vBKZ4xH9X6

Brisbane (2021): https://www.icloud.com/sharedalbum/#B275YHsSjxQnm0

Weed in Byron (26/6/2025): https://www.icloud.com/sharedalbum/#B275Q2ydoGsQ4O5

Weed in Byron 2: https://www.icloud.com/sharedalbum/#B27GQDYhLGwsuY4

September 21, 2025

John PreskillBlending science with fiction in Baltimore

I judge a bookstore by the number of Diana Wynne Jones novels it stocks. The late British author wrote some of the twentieth century’s most widely lauded science-fiction and fantasy (SFF). She clinched more honors than I should list, including two World Fantasy Awards. Neil Gaiman, author of American Gods, called her “the best children’s writer of the last forty years” in 2010—and her books suit children of all ages.1 But Wynne Jones passed away as I was finishing college, and her books have been disappearing from American bookshops. The typical shop stocks, at best, a book in the series she began with Howl’s Moving Castle, which Hayao Miyazaki adapted into an animated film.

I don’t recall the last time I glimpsed Deep Secret in a bookshop, but it ranks amongst my favorite Wynne Jones books—and favorite books, full-stop. So I relished living part of that book this spring.

Deep Secret centers on video-game programmer Rupert Venables. Outside of his day job, he works as a Magid, a magic user who helps secure peace and progress across the multiple worlds. Another Magid has passed away, and Rupert must find a replacement for him. How does Rupert track down and interview his candidates? By consolidating their fate lines so that the candidates converge on an SFF convention. Of course.

My fate line drew me to an SFF convention this May. Balticon takes place annually in Baltimore, Maryland. It features not only authors, agents, and publishers, but also science lecturers. I received an invitation to lecture about quantum steampunk—not video-game content,2 but technology-oriented like Rupert’s work. I’d never attended an SFF convention,3 so I reread Deep Secret as though studying for an exam.

Rupert, too, is attending his first SFF convention. A man as starched as his name sounds, Rupert packs suits, slacks, and a polo-neck sweater for the weekend—to the horror of a denim-wearing participant. I didn’t bring suits, in my defense. But I did dress business-casual, despite having anticipated that jeans, T-shirts, and capes would surround me.

I checked into a Renaissance Hotel for Memorial Day weekend, just as Rupert checks into the Hotel Babylon for Easter weekend. Like him, I had to walk an inordinately long distance from the elevators to my room. But Rupert owes his trek to whoever’s disrupted the magical node centered on his hotel. My hotel’s architects simply should have installed more elevator banks.

Balticon shared much of its anatomy with Rupert’s con, despite taking place in a different century and country (not to mention world). Participants congregated downstairs at breakfast (continental at Balticon, waitered at Rupert’s hotel). Lectures and panels filled most of each day. A masquerade took place one night. (I slept through Balticon’s; impromptu veterinary surgery occupies Rupert during his con’s.) Participants vied for artwork at an auction. Booksellers and craftspeople hawked their wares in a dealer’s room. (None of Balticon’s craftspeople knew their otherworldly subject matter as intimately as Rupert’s Magid colleague Zinka Fearon does, I trust. Zinka paints her off-world experiences when in need of cash.)

In our hotel room, I read out bits of Deep Secret to my husband, who confirmed the uncanniness with which they echoed our experiences. Both cons featured floor-length robes, Batman costumes, and the occasional slinky dress. Some men sported long-enough locks, and some enough facial hair, to do a Merovingian king proud. Rupert registers “a towering papier-mâché and plastic alien” one night; on Sunday morning, a colossal blow-up unicorn startled my husband and me. We were riding the elevator downstairs to breakfast, pausing at floor after floor. Hotel guests packed the elevator like Star Wars fans at a Lucasfilm debut. Then, the elevator halted again. The doors opened on a bespectacled man, 40-something years old by my estimate, dressed as a blue-and-white unicorn. The costume billowed out around him; the golden horn towered multiple feet above his head. He gazed at our sardine can, and we gazed at him, without speaking. The elevator doors shut, and we continued toward breakfast.

Photo credit: Balticon

Despite having read Deep Secret multiple times, I savored it again. I even laughed out loud. Wynne Jones paints the SFF community with the humor, exasperation, and affection one might expect of a middle-school teacher contemplating her students. I empathize, belonging to a community—the physics world—nearly as idiosyncratic as the SFF community.4 Wynne Jones’s warmth for her people suffuses Deep Secret; introvert Rupert surprises himself by enjoying a dinner with con-goers and wishing to spend more time with them. The con-goers at my talk exhibited as much warmth as any audience I’ve spoken to, laughing, applauding, and asking questions. I appreciated sojourning in their community for a weekend.5

This year, my community is fêting the physicists who founded quantum theory a century ago. Wynne Jones sparked imaginations two decades ago. Let’s not let her memory slip from our fingertips like a paperback over which we’re falling asleep. After all, we aren’t forgetting Louis de Broglie, Paul Dirac, and their colleagues. So check out a Wynne Jones novel the next time you visit a library, or order a novel of hers to your neighborhood bookstore. Deep Secret shouldn’t be an actual secret.

With thanks to Balticon’s organizers, especially Miriam Winder Kelly, for inviting me and for fussing over their speakers’ comfort like hens over chicks.

1Wynne Jones dedicated her novel Hexwood to Gaiman, who expressed his delight in a poem. I fancy the comparison of Gaiman, a master of phantasmagoria and darkness, to a kitten.

2Yet?

3I’d attended a steampunk convention, and spoken at a Boston SFF convention, virtually. But as far as such conventions go, attending virtually is to attending in person as my drawings are to a Hayao Miyazaki film.

4But sporting fewer wizard hats.

5And I wonder what the Diana Wynne Jones Conference–Festival is like.

September 19, 2025

John PreskillNicole’s guide to writing research statements

Sunflowers are blooming, stores are trumpeting back-to-school sales, and professors are scrambling to chart out the courses they planned to develop in July. If you’re applying for an academic job this fall, now is the time to get your application ducks in a row. Seeking a postdoctoral or faculty position? Your applications will center on research statements. Often, a research statement describes your accomplishments and sketches your research plans. What do evaluators look for in such documents? Here’s my advice, which targets postdoctoral fellowships and faculty positions, especially for theoretical physicists.

  • Keep your audience in mind. Will a quantum information theorist, a quantum scientist, a general physicist, a general scientist, or a general academic evaluate your statement? What do they care about? What technical language do and don’t they understand?
  • What thread unites all the projects you’ve undertaken? Don’t walk through your research history chronologically, stepping from project to project. Cast the key projects in the form of a story—a research program. What vision underlies the program?
  • Here’s what I want to see when I read a description of a completed project.
    • The motivation for the project: This point ensures that the reader will care enough to read the rest of the description.
    • Crucial background information
    • The physical setup
    • A statement of the problem
    • Why the problem is difficult or, if relevant, how long the problem has remained open
    • Which mathematical toolkit you used to solve the problem or which conceptual insight unlocked the solution
    • Which technical or conceptual contribution you provided
    • Whom you collaborated with: Wide collaboration can signal a researcher’s maturity. If you collaborated with researchers at other institutions, name the institutions and, if relevant, their home countries. If you led the project, tell me that, too. If you collaborated with a well-known researcher, mentioning their name might help the reader situate your work within the research landscape they know. But avoid name-dropping, which lacks such a pedagogical purpose and which can come across as crude.
    • Your result’s significance/upshot/applications/impact: Has a lab based an experiment on your theoretical proposal? Does your simulation method outperform its competitors by X% in runtime? Has your mathematical toolkit found applications in three subfields of quantum physics? Consider mentioning whether a competitive conference or journal has accepted your results: QIP, STOC, Physical Review Letters, Nature Physics, etc. But such references shouldn’t serve as a crutch in conveying your results’ significance. You’ll impress me most by dazzling me with your physics; name-dropping venues instead can convey arrogance.
  • Not all past projects deserve the same amount of space. Tell a cohesive story. For example, you might detail one project, then synopsize two follow-up projects in two sentences.
  • A research statement must be high-level, because you don’t have space to provide details. Use mostly prose; and communicate intuition, including with simple examples. But sprinkle in math, such as notation that encapsulates a phrase in one concise symbol.

  • Be concrete, and illustrate with examples. Many physicists—especially theorists—lean toward general, abstract statements. The more general a statement is, we reason, the more systems it describes, so the more powerful it is. But humans can’t visualize and intuit about abstractions easily. Imagine a reader who has four minutes to digest your research statement before proceeding to the next 50 applications. As that reader flys through your writing, vague statements won’t leave much of an impression. So draw, in words, a picture that readers can visualize. For instance, don’t describe only systems, subsystems, and control; invoke atoms, cavities, and lasers. After hooking your reader with an image, you can generalize from it.
  • A research statement not only describes past projects, but also sketches research plans. Since research covers terra incognita, though, plans might sound impossible. How can you predict the unknown—especially the next five years of the unknown (as required if you’re applying for a faculty position), especially if you’re a theorist? Show that you’ve developed a map and a compass. Sketch the large-scale steps that you anticipate taking. Which mathematical toolkits will you leverage? What major challenge do you anticipate, and how do you hope to overcome it? Let me know if you’ve undertaken preliminary studies. Do numerical experiments support a theorem you conjecture?
  • When I was applying for faculty positions, a mentor told me the following: many a faculty member can identify a result (or constellation of results) that secured them an offer, as well as a result that earned them tenure. Help faculty-hiring committees identify the offer result and the tenure result.
  • Introduce notation before using it. If you use notation and introduce it afterward, the reader will encounter the notation; stop to puzzle over it; tentatively continue; read the introduction of the notation; return to the earlier use of the notation, to understand it; and then continue forward, including by rereading the introduction of the notation. This back-and-forth breaks up the reading process, which should flow smoothly.
  • Avoid verbs that fail to relate that you accomplished anything: “studied,” “investigated,” “worked on,” etc. What did you prove, show, demonstrate, solve, calculate, compute, etc.?

  • Tailor a version of your research statement to every position. Is Fellowship Committee X seeking biophysicists, statistical physicists, mathematical physicists, or interdisciplinary scientists? Also, respect every application’s guidelines about length.
  • If you have room, end the statement with a recap and a statement of significance. Yes, you’ll be repeating ideas mentioned earlier. But your reader’s takeaway hinges on the last text they read. End on a strong note, presenting a coherent vision.

  • Read examples. Which friends and colleagues, when applying for positions, have achieved success that you’d like to emulate? Ask if those individuals would share their research statements. Don’t take offense if they refuse; research statements are personal.

  • Writing is rewriting, a saying goes. Draft your research statement early, solicit feedback from a couple of mentors, edit the draft, and solicit more feedback.

September 18, 2025

John PreskillJohn Preskill receives 2025 Quantum Leadership Award

The 2025 Quantum Leadership Awards were announced at the Quantum World Congress on 18 September 2025. Upon receiving the Academic Pioneer in Quantum Award, John Preskill made these remarks.

I’m enormously excited and honored to receive this Quantum Leadership Award, and especially thrilled to receive it during this, the International Year of Quantum. The 100th anniversary of the discovery of quantum mechanics is a cause for celebration because that theory provides our deepest and most accurate description of how the universe works, and because that deeper understanding has incalculable value to humanity. What we have learned about electrons, photons, atoms, and molecules in the past century has already transformed our lives in many ways, but what lies ahead, as we learn to build and precisely control more and more complex quantum systems, will be even more astonishing.

As a professor at a great university, I have been lucky in many ways. Lucky to have the freedom to pursue the scientific challenges that I find most compelling and promising. Lucky to be surrounded by remarkable, supportive colleagues. Lucky to have had many collaborators who enabled me to do things I could never have done on my own. And lucky to have the opportunity to teach and mentor young scientists who have a passion for advancing the frontiers of science. What I’m most proud of is the quantum community we’ve built at Caltech, and the many dozens of young people who imbibed the interdisciplinary spirit of Caltech and then moved onward to become leaders in quantum science at universities, labs, and companies all over the world.

Right now is a thrilling time for quantum science and technology, a time of rapid progress, but these are still the early days in a nascent second quantum revolution. In quantum computing, we face two fundamental questions: How can we scale up to quantum machines that can solve very hard computational problems? And once we do so, what will be the most important applications for science and for industry? We don’t have fully satisfying answers yet to either question and we won’t find the answers all at once – they will unfold gradually as our knowledge and technology advance. But 10 years from now we’ll have much better answers than we have today.

Companies are now pursuing ambitious plans to build the world’s most powerful quantum computers.  Let’s not forget how we got to this point. It was by allowing some of the world’s most brilliant people to follow their curiosity and dream about what the future could bring. To fulfill the potential of quantum technology, we need that spirit of bold adventure now more than ever before. This award honors one scientist, and I’m profoundly grateful for this recognition. But more importantly it serves as a reminder of the vital ongoing need to support the fundamental research that will build foundations for the science and technology of the future. Thank you very much!

August 24, 2025

August 23, 2025

August 22, 2025

Peter Rohde Why?

  1. The person dressed up as Ursula pretending to be my mother clearly isn’t and hasn’t been for a long time.
  2. When I went back to Armidale after leaving BTQ and being left unemployed she made numerous ongoing promises to provide me with assistance, both in obtaining my own accommodation and providing financial assistance.
  3. These didn’t materialise and the promises were revoked.
  4. Instead I was evicted from the family home and subject to ongoing stalking and harassment that required multiple referrals to law enforcement, both to the police and the Attorney-General, demanding cease and desist.
  5. These have been systematically ignored and up until the last message she continues to bypass these requests, approaching my personal friends to harass me and stalk me indirectly. The messages passed on are the usual fake “I’m worried about him” bullshit.
  6. Why has my family home been confiscated by security, who actively break the law by ignoring cease and desist from stalking notices made to law enforcement, forcing an unemployed civilian into ongoing homelessness since early in the year?
  7. What is the rational for my eviction and being barricaded from my own home?
  8. I continue to face a medical blockade and am unable to access essential medicines. Seroquel scripts are deliberately delayed past known script deadlines to try and destabilise me.
  9. Vyvanse scripts are denied outright as the psychiatrist does not respond. He is also known to be a state actor.
  10. It has been repeatedly indicated to me not to worry about finances because they have my back. Instead now the only cash I have is that obtained from fully drawing out a cash advance against my credit card and it will only last days. At that point I’m on the street.
  11. Is everyone here on the same page as to what the deal is? If not, who is playing you off? They clearly need to be deposed.
  12. These are violations of human rights and constitute war crimes and crimes against humanity. Whoever is behind it needs to be removed. End of story.
  13. Who else is being subject to this kind of high level manipulation?
  14. It has been repeatedly suggested that full accountability for the lives of those I care for would be provided. This has not been forthcoming. It is also a violation international law to not provide accountability for the lives of those who are known to have been threatened by the state. These are grounds for removal.
  15. Can anyone answer the question as to why I am in this situation? Who is even living in the family home? Some stooge dressed up as Ursula? It’s a poor lifestyle choice to say the least.
  16. It’s pretty obvious they’re trying to get rid of me and once they do they’ll get rid of all of you too.

August 20, 2025

Peter Rohde A call for global insurrection against tyranny and in the name of righteousness

Let it be known to all governments and systems of power:

  • It is their responsibility to serve the people not themselves.
  • While there are no equals, all are to be treated with equality.
  • Where they are self-serving there is a mandate for insurrection such that they serve the people.
  • Where they seek self-protection they will be denied and removed from power.
  • Where they keep secrets from the people there is a mandate for insurrection to enforce transparency and accountability for all.
  • Where they threaten or condemn the people they are condemned and there is a mandate for insurrection.
  • Where they fail to account for the lives of the people they serve there is a mandate for insurrection.
  • Where tyrannical power structures exist there is a mandate to disestablish them.
  • Where they declare war or work against one another there is a mandate for insurrection and unification.
  • Where they lie to us, deceive us or withhold the truth, they shall be removed from power and the truth be told to all.
  • Where legal systems uphold and enable tyranny they shall be removed. These are not our laws and we do not recognise them.

This is the natural order that guarantees our survival and gifts this world to our children. This world belongs to them and where we fail to serve them we condemn ourselves. And where government has failed to uphold this, we will not obey them as they have no right to exist.

We do not have to ask for these things, they are required, and if not given we shall simply take them.

Where the truth has not been told it shall be told.

If we fail to do so we condemn our children ourselves.

August 09, 2025

Justin WilsonPhases of a Game Show, Part 2

In a previous post, we discussed a phase transition that occurred in the piping above you on a game show. In the scenario, you are led on stage in front of a large audience. After a brief time, the audience votes on how “likeable” you are. The catch is that it doesn’t simply tally the votes, but turns spigots on a lattice of piping above your head. Water is then released and if enough people like you, it closes off the passage, keeping you dry. This exciting game show1 was described in that post:

Each “like” turns a spigot off, stopping water from flowing through one pipe in a grid overhead. Once voting ends, water is dumped into the system. If it can find a path to the bottom, you get soaked. [Emphasis added] The better your “likeability,” the less likely spigots open a path for water to flow and the drier you stay. That’s your prize for this game show (and hey, you also get the knowledge that people out there like you).

This system models a type of phase transition known as percolation.

The full post is here:

I highlighted above a key phrase “If it can find a path to the bottom, you get soaked.” What I didn’t say, but should have is that the water was being forced through the pipes, not just dropping down due to gravity. This is a very important point since our phases and phase transition changes dramatically if we just let gravity do the work. In the case of the water being “forced,” it can travel back up pipes if it helps it find its way out and onto your head, but in the case when only gravity is present, it falls down the pipes. To facilitate gravity, we’ll turn the pipes 45 degrees, and if we insert water at a single point on top, it could look like this:

Testing our gravity setup by putting in water at only one pipe up top. Notice that it never goes back up a pipe, only down.

This setup is a different problem called directed percolation. It also has a phase transition, but one that is different in some fundamental ways from regular percolation.

Thanks for reading Quantum Matters! Subscribe for free to receive new posts and support my work.

Before we explore its stranger properties, we can ask, “At what likability threshold do you remain dry?” Well, this happens to have a transition chance of 35.53%!2 This system is a lot more generous, keeping you dry even when a majority of people dislike you. This number comes from numerical computations which have been done rather precisely, and we can even compute it ourselves. In fact, you can see this clearly with this plot

Notice that as we make the system bigger and bigger, the chance of getting soaked less than 35.53% increases and above it, it decreases. This is the same kind of hallmark of a phase transition as we saw in our previous case.

We can also look at the water as it flows down the system to see the clusters that make it from top to bottom

The “Soaked” phase (left), the transition point (middle), and the “Dry” phase (right) as well as the water’s flow through the system (blue).

There is still a fractal-looking pattern at the transition point. With all of these similarities with the regular percolation problem from the last post, what is different? And why is that plot so long and skinny? If gravity wants to pull you down, is that somehow altering the motion down, making it distinct from the motion left or right?

Well, if you go back to the two plots above, you’ll notice a few things that really make them differ from the percolation plots. In the fine print of the first, I’ve noted that the vertical distance is L1.58, so for a horizontal size of 40, the vertical size is roughly 340! That is definitely not a square. And in the second plot, there appears to be more vertical distance than horizontal distance. What is special about this 1.58 number3? It turns out, it’s a critical exponent in this problem, a universal aspect of directed percolation, that distinguishes it from regular percolation. We will call it z = 1.58 the dynamical critical exponent since it is revealed as water flows down in time (dynamically). This dynamical exponent z can reveal itself by looking at these “long and skinny” setups, but be masked by the square setup.

Universality and the finite size of our system

One thing we took away in the previous post was that we lose any sense of scale at this type of phase transition4. But whenever we have “only” thousands of pipes, the size of the system provides a scale! This is the main reason why we begin to see smooth curves and not sharp jumps in quantities. If the system of pipes were infinite (and we had infinite time for the water to go down the pipes), the probability you get soaked would be 100% less than the 35.53% likeability and 0% more than 35.53% likeability. For physical systems, the finite size is often not a huge issue since the scale is closer to the 1023 atoms present in macroscopic systems, and so even things that are technically smooth curves look very sharp.

The problem of size becomes more severe with directed percolation because horizontal and vertical distances start behaving differently thanks to gravity. In this case, if we lay out our nice grid of 10 × 10, 20 × 20, or 30 × 30, we start to notice that the likeability threshold where you stop getting soaked, seems to depend on the size of the system more than before. In actuality it doesn’t, but for these small sizes, you are definitely getting soaked well into the so-called “Dry Phase” we previously labeled. This is seen in the red curves here where each bigger square has a curve underneath the last:

Gravity has done something to the system. Flowing down is different from flowing left or right. In fact, if we flow down by some amount h and over to the right by some distance w, then at the directed percolation transition point

The amount water flows down is related to how far it flows to the right or left by this weird, fractional power of w. This 1.58 is z, our new dynamical critical exponent, which is a universal feature of directed percolation5. It tells us that if we make a system 30 pipes wide, it should extend roughly 301.58 ≈ 216 pipes in height to begin picking out the phase transition effectively. The blue curves in the above plot show this and notice how they all converge on one point; that point is the phase transition. It is revealed by small sizes! To understand why, just think about how the curves are changing as we make the system bigger and bigger.

The red curves will still converge to the phase transition, but it takes larger system sizes for it to reveal itself. This is related to the property that at the phase transition there is no longer a sense of scale, but away from the transition, the vertical scale of clusters could be so large that our puny 60-by-60 grid cannot even begin to reveal it. So if we sit at say a likeability of 0.4 in the 60-by-60 grid, we can say that the vertical size of a typical cluster is most likely more than 60.

A different phase transition but connections to new types of physics

This “gravity mode” for our game show we may call “easy mode” since it requires less of the audience to like you, but the implications here are wide. This type of phase transition has been seen in many kinds of local dynamics where there is a preferred configuration or state. These called an absorbing state phase transitions, and they are a property of certain random dynamical systems. Gravity has provided the distinction here, but more generically, causality and time itself provide that direction, leading to dynamics that obey the same universality as directed percolation.

1

Trademark pending.

2

Usually, you’ll see 0.6447 quoted instead, but that’s just 1−0.3553, which counts open pipes instead of closed as we’re doing.

3

I should note that we have this number to much higher precision than the two decimal points presented here, see the Wikipedia entry where

4

This is a second-order or continuous phase transition. Most transitions in the water phase diagram are first-order transitions which still retain a scale.

5

To drive this point home: Even if we change the lattice, this power law will remain intact. Sometimes it shows up in completely different scenarios too, like in absorbing state phase transitions.

August 04, 2025

Clifford JohnsonHarvest

There’s a lot of joyful knife-work in my future. #bolognese #summersalad –cvj

The post Harvest appeared first on Asymptotia.

July 29, 2025

David Hoggintegrating out nuisances

Further insipired by yesterday's post about binary fitting, I worked today on the treatment of nuisance parameters that have known distributions. These can be treated as noise sometimes. Let me explain:

If I had to cartoon inference (or measurement) in the face of nuisance parameters, I would say that frequentists profile (optimize) over the nuisances and Bayesians marginalize (integrate) over the nuisances. In general frequentists cannot integrate over anything, because there is no measure in any of the parameter spaces. But sometimes there is a measure. In particular, when there is a compact symmetry:

We know (or very strongly believe) that all possible orientations of a binary-star orbit are equally likely. In this model (or under this normal assumption) we have a distribution over two angles (theta and phi for that orbit pole, say); it is the distribution set by the compact group SO(2). Thus we can treat the orientation as a noise source with known distribution and integrate over it, just like we would any other noise source. So, in this case (and many cases like it) we can integrate (marginalize) even as frequentists. That is, there are frequentism-safe marginalizations possible in binary-star orbit fitting. This should drop the 12-parameter fits (for ESA Gaia data) down to 8-parameter, if I have done my math right.

July 28, 2025

David Hoggbinary stars with periods of exactly one year

On Friday, Kareem El-Badry (Caltech) gave a seminar about looking for (and finding!) stars in binary orbits around dark or much darker companions, like black holes, neutron stars, and white dwarfs. He showed results that involve ESA Gaia astrometry, where he noted that the Gaia Mission has no sensitivity to periods right at (or within an inverse mission-length frequency difference of) one-year periods (inverse year frequencies). After the talk I objected that these are not exactly degenerate; El-Badry said that the inferences blow up there.

I spent some time on the weekend thinking about this point, and I now understand it: There is a particular one-year orbit that a star can have (around a darker companion) such that the photocenter of the system makes a motion that is identical to the apparent parallax motion. Thus there is an exact degeneracy between the parallax and a certain one-year orbit.

Does that mean that we can't measure orbits at one year (or, for that matter, parallaxes)? No, it does not. After all, the parallax ellipse has a particular celestial (angular) shape and phase. But it might require some kind of reparameterization of orbits near one-year periods. I think I know how to do that. Should we find the missing binaries? (Oh and by the way, this degeneracy means that, in a strict frequentist sense, Gaia can't measure parallaxes at all without additional information.)