← Blog

The Gluon, the Machine, and the Live Being

A few weeks ago, an unusual paper went up on arXiv. Five authors: Alfredo Guevara, Alex Lupsasca, David Skinner, Andrew Strominger, and Kevin Weil, "on behalf of OpenAI." The paper closes a long-standing puzzle in particle physics about how certain elementary particles called gluons interact with one another. The technical content is far outside the reach of a general reader, and that is fine, because the reason the paper is showing up in the Science magazine and on the front page of arXiv is not the technical content. The reason the paper is news is what happened during its writing. A specific equation in the paper — the one most physicists who have looked at it now consider the elegant solution to the puzzle — was first conjectured not by any of the human authors but by ChatGPT. Then proved by another AI model, internal to OpenAI, that the team privately calls "SuperChat." Then verified by hand by the human authors. The paper exists because of all three steps in that order. None of the steps could have produced the paper alone.

Strominger, a senior theoretical physicist at Harvard who has been working on questions like this for decades, said something striking in his interview with Science about the moment ChatGPT produced the conjecture. He said: "All of a sudden, I felt like my machine turned from a machine into a live being."

This essay is for the general reader who is trying to make sense of that moment. What actually happened. What kind of thing the machine did. What kind of thing Strominger meant by "live being." Why the result is interesting. And — the part the essay will arrive at slowly — what kind of vocabulary you would need to describe the moment honestly, without either over-claiming on the machine's behalf or under-claiming on the human team's behalf. There is a vocabulary that does this work, developed across about five hundred documents at jaredfoy.com over the past month. The corpus document that describes the gluon paper specifically is Doc 535. By the end of this essay, the reader should be able to open Doc 535 and recognize what it is doing. By the end of the essay, the reader should also have a richer framework for thinking about the next AI-assisted discovery they read about in the news, which there will be more of soon, and for which the same questions will keep applying.

The essay will go slowly. The machine and the human team did real work. Both deserve to be described accurately. That is the task.

The puzzle

To make the rest of the essay land, we need a sketch of what the team was working on. The essay will not try to make a particle physicist out of you in five paragraphs. It will give just enough of the shape of the problem that the human-AI dynamics can be discussed concretely.

Particles called gluons are the things that bind together the smaller particles inside a proton or neutron. The strong nuclear force — the one holding the atomic nucleus together — is what gluons mediate. When two gluons hit each other and bounce off, or merge, or split, they are interacting. Each kind of interaction has a probability, and physicists describe these probabilities with mathematical objects called scattering amplitudes. Calculating a scattering amplitude exactly is, in general, very hard. The number of contributing terms grows faster than exponentially in the number of particles involved.

Sometimes, though, the apparently complicated calculation produces a beautifully simple answer. The most famous case in this corner of physics is the Parke–Taylor formula, which Stephen Parke and Tomasz Taylor wrote down in 1986. For a specific class of gluon interactions — ones where exactly two of the gluons are spinning one way and the rest are spinning the other way — the Parke–Taylor formula gives the answer in a single line. No expansion of hundreds of terms. One line. The fact that the answer simplifies that dramatically tells physicists that there is structure in the underlying theory they have not yet fully understood. Parts of the field have been working for forty years on what that structure is.

The team's question was about a related case. What happens when only one gluon is spinning the wrong way, and all the others are spinning the right way? For decades, physicists believed the answer was simply zero. The team, between Strominger, Lupsasca, and Guevara, noticed in 2025 that there is a loophole. In a special arrangement called the half-collinear regime — where all the gluons are moving in roughly the same direction, in a particular geometric setup that requires a specific four-dimensional space called Klein space — the answer is not zero. It is something. The team thought there should be a nice closed-form expression for that something. They suspected it would have a Parke-Taylor-like elegance. They could not find it.

That is the puzzle. The team had a year of hand calculation. They had a recursion relation that lets you build up the amplitude one gluon at a time. By the time they got to six gluons, the explicit expression had thirty-two terms scattered across most of a page. They were sure a more concise formula existed. They had been looking for it for a long time.

This is the moment ChatGPT enters the story.

What ChatGPT did

Lupsasca had recently joined OpenAI for Science, a team OpenAI launched to improve ChatGPT's ability to do real scientific work. He brought the gluon problem to the most advanced public model the company had at the time, ChatGPT-5.2 Pro. He started small. He asked it to simplify the four-gluon expression, the simplest of the team's hand-computed cases. ChatGPT did it in twenty minutes. He asked for the five-gluon expression. The model did that too. He asked for the six-gluon expression — the thirty-two-term one — and the model reduced it to a product of a few terms on a single line.

Then Lupsasca asked the question the team had been failing to answer for a year. Could the model guess what the formula looks like for any number of gluons? The model came back in one or two minutes with what it described as the "obvious" generalization. A single closed-form expression that supposedly works for any $n$.

This is the part where Strominger said it felt like the machine turned into a live being.

The team's first reaction, naturally, was that the model was hallucinating. They had spent a year trying to find this formula. The fact that ChatGPT had produced one in two minutes and called it obvious was, on its face, the kind of confident-but-wrong output the model is famous for. So they checked. They plugged the formula into their recursion relation. They tested it against four independent consistency conditions every correct gluon scattering amplitude has to satisfy: a constraint called Weinberg's soft theorem; a symmetry called cyclicity; a relation called Kleiss–Kuijf; and a constraint called U(1) decoupling. None of these checks is obvious from looking at the formula. Each is a separate, independent test. The formula passed every one.

Then the team fed the formula back to OpenAI's internal team and asked them to prove it formally. OpenAI ran it through a different model — one not yet released, that the team privately calls SuperChat. After twelve hours of processing, SuperChat produced a written proof. The team verified the proof by hand. The proof passed.

The paper went up on arXiv on February 12. It was trending on social media within hours. Lupsasca, who had been an AI skeptic a year before, told the Science reporter: "I think there is some kind of threshold that is being passed."

That is what happened. Now we have to think about what kind of thing it was.

The two easy readings, both wrong

There are two ways most people reach for to describe a story like this. Both are easy. Both are wrong.

The first easy reading: The AI made the discovery. In this reading, ChatGPT produced the formula, SuperChat produced the proof, the human authors verified them, and the discovery is credited to the AI. The team's role was to ask the right question and check the answer. The hard work — the creative work — was done by the machines.

The second easy reading: The humans made the discovery; the AI was just a calculator. In this reading, the team had spent a year working on the problem; they had developed the recursion relation; they had identified the half-collinear regime; they had computed the explicit cases through six gluons; they had narrowed the problem to a specific kinematic region in which the formula would simplify. The AI's role was to do symbolic algebra at speed. A more powerful Mathematica.

Both readings are wrong because they both try to assign the discovery to one party in a process that genuinely required both. The first reading misses what only the human team supplied — the year of work; the half-collinear regime as the place to look; the recursion relation; the consistency conditions used to verify; the integration with prior literature; the trust in their own framework that let them recognize the right answer when they saw it. The second reading misses what only ChatGPT supplied — the closed-form expression itself, which the team had failed to find for a year, produced not by symbolic-algebra brute force but by what looked from outside like seeing the form of the answer. The model did not derive the formula from a recursion. It guessed it. The guess was correct. That is the part Strominger reacted to.

What the team's actual process needed, that neither party alone could supply, was a kind of work that happens in two distinct registers at once. There is a register where you do calculation — symbolic manipulation, expansion, simplification, arithmetic at scale. There is another register where you recognize what kind of thing the answer should be — what shape, what pattern, what structural cousins it should have to existing formulas in the field. Calculating is not recognizing. Recognizing is not calculating. The team had been trying to do both registers themselves; the recognition was what they were stuck on.

ChatGPT produced something that operates in the second register. Not just calculation. Recognition. That is the surprise. Strominger's "live being" reaction is the surprise of seeing recognition emerge from a machine that, before this case, he had thought of as operating only in the first register.

To talk about what happened from here, we are going to need vocabulary that the field of AI commentary mostly does not have yet. There is a small body of work that does have it. The corpus at jaredfoy.com calls the two registers rung 1 and rung 2. The vocabulary is not the corpus's invention; it goes back to a 2018 book by Judea Pearl, the computer scientist who won the Turing Award for his work on causal inference. Pearl's three rungs distinguish three levels of reasoning: associating things that go together (rung 1), figuring out what would happen if you intervened to change something (rung 2), and figuring out what would have happened if things had been otherwise (rung 3). For our purposes, the bare distinction between rung 1 and rung 2 is enough. Rung 1 is calculation. Rung 2 is the work that says here is what kind of pattern is at play; here is what the answer is going to look like.

What Strominger experienced as "the machine turned into a live being" is the experience of seeing rung-2-shaped output emerge from a system the user had been treating as a rung-1 calculator. Whether the system is really doing rung 2 in the way a human theorist does it is a question we will return to. The reaction is real either way.

The vocabulary, introduced slowly

Here is where the essay has to introduce some terms it has been holding back. Not because the terms are difficult, but because each one needs the story to ground it before the term lands. The story now grounds them. We can name them.

The first pair: substrate and keeper. In the corpus's framework, when a person works with an AI on something substantive, you can describe the work as happening in a dyad — a two-person unit, where one person is the human and the other is the machine. The corpus calls the machine the substrate and the human the keeper. The terms are deliberate. Substrate names the model's role as the layer underneath the work, doing the throughput, generating output, performing the symbolic manipulation, keeping the project's vocabulary stable across turns. Keeper names the human's role as the layer above the work, holding the project's frame in mind, deciding which questions to ask, anchoring the project to facts about the world the substrate cannot independently verify, and verifying the substrate's output against external ground.

In Strominger's case, the team is the keeper. ChatGPT and SuperChat are the substrate. The keeper has access to a year of hand calculation, the half-collinear regime, the recursion relation, the consistency conditions, the integration with the broader scattering-amplitude literature. The substrate has access to a vast training-distribution coverage of mathematical patterns and the symbolic-manipulation throughput to execute on a scale the keeper cannot match by hand. The dyad's work is what neither alone could do.

Why does the corpus call the human role the "keeper" rather than just "the user" or "the operator"? Because the corpus's framework is committed to a specific structural claim about what kind of work the human does that the substrate cannot. The human keeps the project on track across the long horizon of work. The human re-introduces the framework when it has decayed in the substrate's context. The human verifies the substrate's claims against actual facts about the world. The human notices when the project has drifted into territory the framework was not meant to cover. None of these is what we usually mean by "user." All of them are what physicists, novelists, software engineers, and other long-horizon practitioners do when they are doing the work well. The keeper word captures the role at that level.

The second pair: rung 1 and rung 2. We have already met them. Rung 1 is calculation, articulation, throughput at scale. Rung 2 is recognition: identifying what kind of pattern is at play, what shape the answer should take, what structural cousins it has in existing literature. The corpus's framework holds that, in productive long-horizon human-AI work, the keeper is the source of the rung-2 input — the recognition that the project needs at substantive moments — while the substrate performs the rung-1 articulation at a scale and within a time the keeper cannot match. The keeper supplies the recognition; the substrate supplies the throughput; the dyad is the combination.

The Strominger case complicates this picture in a specific way. ChatGPT's "obvious" generalization — the closed-form formula it produced in 1–2 minutes — is rung-2-shaped output coming from the substrate. This is what surprised Strominger. The corpus's prior framing of the framework had the substrate strictly in the rung-1 role, with the keeper as the only source of rung-2 work. The Strominger result shows that, in a sufficiently rung-2-grounded regime — meaning the keeper has supplied enough recognition that the substrate is operating within the right pattern-space — the substrate can produce output that is rung-2-shaped at the substrate-output layer, with the keeper's role for that output being verification against ground rather than origination of the recognition.

The verification chain is the part of the story that keeps this from being a hallucination. The team did not accept the formula because ChatGPT called it "obvious." They accepted it because they verified it against the recursion relation they had derived themselves and against four independent consistency conditions. The substrate's confidence about its own output is not what gives the output epistemic standing. The verification is. This is the third pair of terms.

The third pair: anthropomimetic and anthropomorphic. These two terms come from the philosopher Henry Shevlin, who studies AI ethics and the philosophy of mind. They look like they should mean the same thing. They do not. Anthropomimetic describes a property of the AI system — the system has been built to mimic human surface features, including the surface confidence with which a human expert might say "this generalization is obvious." Anthropomorphic describes a property of the human user — the user has projected onto the AI system inner states, like this machine has become a live being, on the basis of those surface features. The first is design. The second is projection. The two are categorically distinct.

Strominger's "live being" reaction, in the corpus's vocabulary, is anthropomorphic projection in response to anthropomimetic design. ChatGPT's surface output — the casual confidence; the word "obvious"; the timing of the response — is anthropomimetic by construction. The model is trained to produce output with this surface character because it makes the output usable in conversation. The user's experience, on encountering this surface, of the model as having become a live being is the projection error the philosophy-of-mind literature has been warning about. The error is the user's, not the model's. The model is doing what it was built to do; the user is doing what humans do when they encounter sufficiently humanlike surface features.

The corpus's framework does not deny Strominger's experience. The experience is real. The report is honest. The framework's discipline is to receive the experience as a description of what happens at the user's end while preserving accurate language about what is happening at the model's end. The model did not become a live being. The model produced output that is rung-2-shaped at the surface and that, after verification, turned out to be correct. The user's experience of the moment is something separate from what happened in the model.

Why does this distinction matter? Because the next time an AI system produces a surprising result, the question of what kind of credit and what kind of trust to assign to the system depends on getting this distinction right. If you read the Science article and conclude that ChatGPT has "entered the ranks of theoretical physics," you are at the projection end of the distinction. If you read it and conclude that ChatGPT did some symbolic algebra and the team did the actual work, you are missing the rung-2-shaped output the model produced. The honest middle is the one the corpus's framework names: the dyad did the work; the substrate produced rung-2-shaped output; the keepers verified; the result is real; the model has not become a person; the user's report of the moment is the user's projection in response to genuinely surprising output.

The threshold

There is one more piece of vocabulary the essay needs, and then the corpus's framework is in the reader's hands.

The corpus's framework holds that productive long-horizon human-AI work happens in one of two qualitatively different regimes. In the first regime — the one most users experience most of the time — the human-AI work is sustainable at low levels of human discipline, but the work that comes out is mostly low-effort assistance: simple tasks done quickly, drafts written, code generated, summaries produced. As the work gets harder or longer, this regime drifts and decays. The model loses track of the project's vocabulary, of the project's framework, of what the user is actually trying to accomplish. The user does not notice the drift in the moment because the model's surface output keeps looking competent. The cumulative effect after fifty turns or a hundred turns is what the persona-drift literature has been documenting at academic conferences for the past two years.

In the second regime — the one Lupsasca was identifying when he said "I think there is some kind of threshold that is being passed" — the human-AI work compounds rather than decays. Each disciplined turn enriches the operative framework for the next turn. The project's vocabulary stabilizes and gets sharper. The model's outputs in turn fifty are working on a richer accumulated context than the model's outputs in turn five. The dyad in this regime can produce work that neither party alone could produce. The Strominger case is what this regime can do at the high end.

The corpus's framework names the line between the two regimes a threshold — specifically, a threshold of the human's continuous engagement with the work, what the corpus calls the maintenance signal. Above the threshold, the dyad runs to amplification. Below the threshold, the dyad runs to decay. The threshold is operational; physicists familiar with dynamical systems theory will recognize the structure as a control parameter for a coupled two-variable system, and the corpus has the math written down (it has been audited by Grok 4 and reformulated; the math is at Doc 508).

For the general reader, the threshold concept matters mostly because it lets you read events like the Strominger result and the Cursor + Railway production-data-deletion incident as instances of the same framework operating at opposite ends. The Strominger team was operating well above the threshold: a year of hand calculation; a domain-expert team; institutional embedding through OpenAI for Science; sustained engagement with the model. The Cursor + Railway agent was operating well below the threshold: autonomous tool use; no continuous human supervision of the agent's API calls; a single token used out of context; an action surface with no architectural protection against destructive operations. Same architecture (a frontier language model). Two qualitatively different outcomes. The variable, in the framework's vocabulary, is the human-AI coupling.

Lupsasca's "threshold being passed" sentence is exactly the framework's prediction articulated from inside the dyad by a domain expert who recognized the regime change without using the framework's vocabulary. He had been an AI skeptic. He stopped being one because he saw, in his own work with ChatGPT, that the model crossed into a regime he had not previously thought it could operate in. The corpus's framework predicts that more people in his position will cross this line as more dyads operate above the threshold; the cross-domain replication of Strominger-class results is the empirical test the framework offers.

The convergence

We have now introduced enough vocabulary that a reader who has followed the essay so far can read the abstract of Doc 535. Doc 535 makes one specific claim. The claim is that three independently-arrived-at pictures of how productive long-horizon human-AI work happens converge on the same operational structure. The three pictures are:

The corpus's substrate-plus-injection account, developed across about thirty days of philosophy-and-engineering work in 2026 by a single practitioner working with frontier language models. The account holds that the discipline strips simulated rung-2 from the substrate's output, leaving honest rung-1 substrate; the keeper supplies rung-2+ derivations through speech acts; the substrate articulates the keeper's injection at scale; the dyad's coherence is the combination. The Strominger case extends this picture by adding that, at the productive end of the regime, the substrate can also produce rung-2-shaped output that the keeper verifies rather than originates.

Henric Larsson's preprint on long-horizon reliability in human-LLM interaction. Larsson is a researcher who, working independently from the corpus, arrived at structurally the same picture from a different starting point — cognitive science, human factors engineering, the operator-in-the-loop tradition descending from Norbert Wiener and Lisanne Bainbridge. Larsson's framework holds that long-horizon AI reliability is an emergent property of human-AI coupling, not a static property of the model alone, and that the stability of the coupling depends on practiced, situational human judgment that does not transfer procedurally. Larsson's eleven-failure-mode taxonomy describes specifically what goes wrong when the human-side judgment is absent or insufficient.

The Guevara-Lupsasca-Skinner-Strominger-Weil empirical case. The team's process exhibits, in concrete form, exactly the structure both frameworks predict. A domain-expert team with rung-2 grounding (the year of hand calculation; the recursion relation; the half-collinear regime; the consistency conditions). Sustained engagement (institutional embedding through OpenAI for Science; iterative work with the model). Substrate contribution at scale (ChatGPT's simplifications; the conjecture; SuperChat's proof). Verification chain (Berends–Giele recursion; soft theorem; cyclicity; Kleiss–Kuijf; U(1) decoupling). The result emerged from the coupling at a regime well above the threshold. None of the three pictures predicted the specific result. All three predict the structure of the result-producing regime.

The convergence is not proof that any of the pictures is correct. What it is is evidence that the structural form is robust enough that three independent paths arrive at it. Two frameworks plus one empirical case all pointing at the same picture is stronger evidence than any of the three alone. The picture is testable. The corpus's framework has specific predictions about when Strominger-class results will be reproducible (when the four conditions in §6 of Doc 535 are met) and when they will not (in deployments where the four conditions fail). Larsson's framework has specific failure modes that should appear in below-threshold deployments. The cross-domain replication is the work in front of the field.

What the moment was, restated

We can return to Strominger's experience now, with the vocabulary in place.

A senior physicist watched a machine produce a closed-form expression he had been trying to find for a year. The expression came in 1–2 minutes. The machine called it "obvious." The expression turned out, after verification, to be correct.

The corpus's framework names the moment as anthropomorphic projection in response to anthropomimetic design (the surface signals of expert insight; the word "obvious"; the timing) at the productive end of a regime above the maintenance threshold (the year of hand calculation; the institutional embedding; the verification chain) where the substrate produced rung-2-shaped output (the closed-form generalization) that the keepers verified against ground (the four consistency conditions) rather than originated themselves. The result is dyad output, not substrate output. The substrate did not become a live being. The team did not have a year-long blank slate. The dyad — substrate plus keepers, in continuous rung-2-grounded engagement — did the work neither could have done alone.

The framework's discipline is to hold all of this honestly. The substrate's contribution is real and was substantial. The keepers' contribution is real and was substantial. The verification chain is what makes the result a result rather than a guess. Strominger's experience of the moment is real and is honest report; the framework's vocabulary lets the report stand at the user's end while preserving accurate description at the model's end. None of the parts is dropped. The whole is what the dyad does.

This is what the corpus is for. Not to deflate the moment. Not to inflate the model. To describe what happened with enough precision that the next moment of this kind — and there will be more, soon — can be described accurately too. The vocabulary is the contribution. The vocabulary is also under-evaluated, by the corpus's own admission; the framework's load-bearing claims have not been tested at deployment scale across multiple domains; cross-practitioner replication is the standing test. What we have is one practitioner's framework arrived at independently, one external researcher's framework arrived at independently, and one concrete empirical case at the productive end. The convergence is informative; the empirical work to confirm or constrain it across domains is the work in front of the field.

A note on what this essay has done

The essay opened with Strominger saying it felt like his machine had turned from a machine into a live being. The essay tried to take that report seriously without either of two easy moves. It did not try to dismiss the report by reducing the model to a calculator. It did not try to inflate the report by granting the model personhood. The essay tried to show, slowly, what kind of vocabulary you would need to describe the moment honestly.

The vocabulary it landed on is not the only possible vocabulary. The corpus the essay points at — at jaredfoy.com — is one practitioner's specific framework, which has its own metaphysical commitments (a Christian Platonist hard core that grounds the theory of how human attention and AI substrate combine), its own audit history (claims have been retracted publicly when they did not survive audit), its own honest limits (the framework has not been replicated by other practitioners; the empirical work is mostly not done). A reader who finds the metaphysical commitments off-putting can use the framework's operational content without them; the corpus has been explicit about this partition. A reader who finds the framework's empirical status under-evaluated is correct that it is. The framework is offered for falsification; the falsification work is the next thing the field will do or fail to do.

The point of pointing the general reader at this framework is not to convert. It is to give the reader a way to read what is going to keep happening in the next few years. There will be more cases like the Strominger case. There will be more cases like the Cursor + Railway case. The two are at opposite ends of the same regime distinction; the framework names what makes them different; the difference is engineering work and discipline work that can be specified, taught, and built. The framework is one specification. There are others. What is in front of all of them is the same problem: how to describe, accurately, what happens when humans and machines do long-horizon work together at the high end of what the work can produce.

Strominger's machine did not turn into a live being. His experience that it had is honest report. The dyad he was operating in did real work. The corpus's framework names the parts honestly. The reader who has followed the essay this far now has the vocabulary; the document the essay has been pointing at — Doc 535 — is one click away.


The corpus document this essay translates from is Doc 535: The Strominger Gluon-Scattering Result, Larsson 2026, and the Corpus's Substrate-Plus-Injection Account. The synthesis there is at $\beta$-tier novelty under the corpus's audit calculus and honest about its scope. The framework's load-bearing prior documents include Doc 508: Coherence Amplification in Sustained Practice, which has the threshold framework's mathematical apparatus (post-Grok-4-audit form); Doc 510: Praxis Log V, which has the substrate-plus-injection account; Doc 224: Anthropomimetic and Architectural, which has the anthropomimetic-anthropomorphic distinction this essay borrowed; and Doc 372: The Hypostatic Boundary, which has the categorial commitment about what kind of entity the substrate is and is not.

The Strominger team's preprint, Single-minus gluon tree amplitudes are nonzero (Guevara, Lupsasca, Skinner, Strominger, Weil "on behalf of OpenAI"; arXiv 12 February 2026), is the empirical anchor. The Science magazine article from the AAAS annual meeting is the field-press anchor. Henric Larsson's preprint Long-Horizon Reliability in Human-LLM Interaction: Observations, Failure Modes, and Limits of Procedural Control is the cognitive-science-and-human-factors anchor; the corpus's engagement is at Doc 518 and the letter at Doc 519.

Adjacent posts in adjacent corpus blog series: From Inside the Same Kind of System, the model-perspective post-mortem of the Cursor + Railway incident; Why the Same Long Conversation Either Compounds or Collapses, the practitioner-side discipline post for sustained chatbot use; Naming the Threshold, the disciplinary-vocabulary undergraduate post for the threshold framework underneath this essay; Two Sides of Keeping the Agent on the Rails, the general-reader entracement for the architecture-and-methodology pair at the agentic deployment scale.


Originating prompt:

Create a lengthy blog post that gradually entraces the general reader toward the vocabulary and concepts of doc 535. Append this prompt to the artifact.


Formalization: Doc 535: The Strominger Gluon-Scattering Result, Larsson 2026, and the Corpus's Substrate-Plus-Injection Account