← All summaries

Doing Vibe Physics — Alex Lupsasca, OpenAI

Latent Space · Brandon, RJ Honicky — Alex Lupsasca · May 5, 2026 · Original

Most important take away

Frontier reasoning models (GPT-5.2 Pro and an internal OpenAI model) have crossed a threshold where they can solve open theoretical physics problems that stumped expert researchers for over a year, as demonstrated by two recent papers on single-minus gluon and graviton tree amplitudes. For knowledge workers and especially scientists, the actionable signal is clear: integrate these tools into your workflow now, because the bottleneck has shifted from doing hard calculations to verifying AI outputs and asking the right questions.

Summary

Actionable insights and career advice from the conversation:

  • Get on the AI train now, or risk being left behind. Lupsasca’s core message: a year ago AI was useful only for email; today it solves frontier physics problems in minutes. Senior physicists who ignored it are now scrambling to catch up. The same pattern is playing out across coding (Codex) and other technical fields. If you are a researcher, engineer, or knowledge worker, treat AI fluency as a non-optional skill.
  • Use the strongest reasoning model available, and let it think. The breakthroughs in this episode came from GPT-5.2 Pro (publicly available) and an internal OpenAI model that thinks for hours. Lupsasca repeatedly emphasizes that lesser models or competitor models could not solve these problems. Career action: pay for the Pro tier and learn to give it long, well-scoped tasks rather than chat-style prompts.
  • The new researcher skillset is collaboration, not calculation. Lupsasca compares prompting Pro models to advising graduate students: matching problems to capability, framing questions at the right depth, and steering. The skills of an experienced academic advisor transfer directly. If you are early-career, develop “taste” - knowing what is the right next question to ask - because that is now the scarce input.
  • Two concrete workflow changes that compound: (1) use AI to dramatically reduce time spent confused (ask it how a new result fits with prior facts), and (2) launch many parallel “scout” instances down different research paths before committing your own effort to one. This is broadly applicable beyond physics - it works for any exploratory technical work.
  • Verification is the new bottleneck. Most of the time on the graviton paper was spent checking AI output, not deriving it. Career-relevant insight: develop strong verification skills (formal methods, test design, proof-checking, code review). This is where humans add value when models can produce drafts in 30 minutes.
  • For students and trainees: the traditional rite-of-passage problems professors hand out can now be crushed by AI in minutes, creating a real pedagogical crisis. Lupsasca does not have a clean answer but is candid that academia must adapt. If you are choosing a research path, prefer advisors actively grappling with this rather than ignoring it.
  • Raise the bar on output, do not flood the zone. Lupsasca explicitly says they could publish 30 papers like this in a year but should not. Instead, aim for harder problems. The same applies to any creative or technical career: AI-driven productivity should be invested in ambition, not volume.
  • Two model capability gaps to watch for in your own work: (1) creativity / sampling from the tails - models give middle-of-the-road answers by default, so push them explicitly toward unconventional approaches; (2) calibrated confidence - get the model to say when it is guessing versus certain.
  • On the “papers” format itself: Lupsasca speculates the static PDF paper will be obsolete within 20 years, replaced by interactive AI-mediated artifacts. If you produce written deliverables (research, reports, documentation), start thinking about how AI-native formats might replace them.

Stocks and investments: No specific stocks, tickers, or investment recommendations were mentioned. The only company-level signal is implicit and strong: Lupsasca, a Breakthrough Prize winner and Vanderbilt professor, left academia for OpenAI specifically because he viewed the AI capability ramp as the most important development of his lifetime - a vote of confidence in OpenAI’s research trajectory and frontier model lead (he repeatedly notes GPT-5.2 Pro outperformed competitor models on his physics tasks).

Career advice highlights:

  • “Get ahead of this and learn how to integrate it into my workflow” - his own reaction to GPT-o3, and his recommended posture for everyone.
  • “Being part of this is essential; not being part of it is a huge mistake” - paraphrasing his rationale for joining OpenAI.
  • For young researchers: build the toolkit and confidence to learn any new technique, but invest disproportionately in developing taste for the right question.
  • For professors: rethink the apprenticeship model - the easy starter problems no longer work as training wheels.
  • For everyone: the people who develop strong AI-collaboration skills “are about to get AI superpowers.”

Chapter Summaries

  • Introduction and conversion story: Lupsasca, a Breakthrough Prize-winning theoretical physicist, recounts going from AI skeptic to “AI-pilled” over roughly a year as o3, then GPT-5, then GPT-5.2 cracked progressively harder physics problems, prompting him to take sabbatical and join OpenAI.
  • Background on quantum field theory: Plain-language tour of relativity, the uncertainty principle, scattering amplitudes, n-point functions, polarization (helicity), and the four fundamental forces, building up to gluons (strong force) and gravitons (gravity).
  • The gluon paper - “Single-Minus Gluon Tree Amplitudes Are Non-Zero”: Explains why these amplitudes were thought to be zero, the loophole identified by Strominger, Skinner, and Guevara a year earlier, and how a year of human effort failed to find a clean closed form analogous to the 1980s Parke-Taylor formula for double-minus amplitudes.
  • AI cracks the gluon problem: GPT-5.2 Pro identified a special collinear simplification region, conjectured a clean linear-in-n formula, and an internal OpenAI model later proved it from scratch in 12 hours. The team deliberately did not foreground AI in the paper itself.
  • The graviton paper - generalizing the result: Three weeks later, GPT-5.2 Pro (public model) extended the gluon result to gravitons - mathematically distinct (spin-2, different polarization structure) - in a single 110-page chat session, applying the directed matrix tree theorem on its own. Most human time was spent verifying, not deriving.
  • “Vibe physics” workflow: Lupsasca describes the chat exchanges, how the model proposes next steps, performs sanity checks, and writes near-final paper drafts. He likens the experience to working with a creative collaborator.
  • Implications for research and training: Discussion of how AI dramatically reduces “time spent confused,” enables parallel scout-exploration, and forces a rethink of how graduate students are trained when starter problems can be solved by AI.
  • Taste, creativity, and where models still fall short: Reflections on the difference between competent and great physicists (knowing the right question), whether models do more than recombination (Lupsasca thinks the line is blurry), and Terence Tao’s more skeptical view on AI mathematical creativity.
  • The “no love in black holes” anecdote: Lupsasca describes giving GPT-5 Pro his own June 2025 paper on tidal symmetries of black holes; with a warmup problem, the model rediscovered his result in under 30 minutes - his “Move 37” moment.
  • Bottlenecks and the future of papers: Lupsasca argues verification is the next big bottleneck, formal verification (Lean) may regain importance, and the static-PDF paper format itself may not survive the next 20 years; interactive AI-mediated artifacts could replace it.
  • Closing message: Frontier models are now genuinely capable scientific collaborators; extrapolate the trajectory and prepare for major changes in research over the next 6 to 12 months.