← All summaries

20VC: Sam Altman on The Trajectory of Model Capability Improvements: Will Scaling Laws Continue | Semi-Conductor Supply Chains | What Startups Will be Steamrolled by OpenAI and Where is Opportunity

20VC · Harry Stebbings — Sam Altman · November 4, 2024 · Original

Most important take away

Don’t build a startup that patches a current model shortcoming, because OpenAI fully intends to fix those shortcomings with each new generation. Build companies that get better as the models get better, and focus on AI-enabled vertical applications (tutors, lawyers, doctors, CAD engineers) where massive new economic value will be created.

Summary

Actionable insights and tech patterns from the conversation:

Career and founder advice:

  • Align your startup with where models are going, not where they are. If your business depends on plugging a current shortcoming of GPT-4 or o1, the next model generation will likely make you irrelevant. Build instead so that better models make your product better.
  • If Altman were a 23-24 year old starting today, he would build an AI-enabled vertical, e.g., the best AI tutor, AI lawyer, or AI CAD engineer. Pattern: pick a high-value domain and deliver a magical end-to-end experience on top of the models.
  • Inexperience is not inherently low-value. Hire on a high talent bar at any age; don’t categorically prefer young-only or experienced-only. Use young hires for fresh-perspective building, experienced hires for high-stakes complex infrastructure.
  • For hard 51/49 decisions, build a network of 15-20 trusted domain experts you can “phone a friend” for context, rather than one universal advisor.
  • A common leadership trap in hypergrowth: spending all effort on the next 10% rather than the next 10x. Going from $1B to $10B requires structural change, internal communication overhauls, and long-lead planning (compute, office space, org design) you must start far in advance.
  • Personal-development angle: Altman flags product strategy as a current weakness and praises product leaders (like Kevin Weil) for discipline, focus, and saying no, “speaking on behalf of the user” rather than chasing fantastical dreams.

Tech patterns and strategic signals:

  • Reasoning (the o-series) is OpenAI’s most strategically important direction. Expect rapid improvement and major unlocks for scientific work and complex code.
  • Scaling laws will continue holding “for a long time,” per Altman. Bet on continued capability improvement.
  • Multimodal and vision capabilities will progress rapidly under the new inference-time-compute paradigm set by o1.
  • Models are depreciating assets, but the revenue a model generates justifies the training spend if you have a sticky product (like ChatGPT). Too many players are training near-duplicate models without a moat.
  • The frame is shifting from “models” to “systems.” Developers already use multiple models; expect heterogeneous AI everywhere.
  • Agents: Altman’s working definition is “something I can give a long-duration task to with minimal supervision.” The interesting use case isn’t booking one restaurant, it’s massively parallel actions a human cannot do (calling 300 restaurants), or an agent acting like a smart senior co-worker on multi-day projects.
  • Agent pricing may move away from per-seat SaaS toward compute-based pricing (e.g., “give me 10 GPUs working on my problems all the time”).
  • No-code AI app building for non-technical founders is coming, but first wave will be tools that make existing coders more productive.
  • Open source has a real place alongside hosted APIs; both will coexist.
  • Anthropic currently has the edge on coding; developers should expect to use multiple models.
  • Best analogy for AI is not the internet or electricity but the transistor: a physics-level discovery with scaling laws that quietly diffuses everywhere and lifts the whole economy.
  • Top risk on Altman’s mind: not chips specifically, but the fractal complexity of coordinating power, networking, chip supply, research timing, and product readiness simultaneously.
  • Five-year outlook: rapid, almost unbelievable scientific and AI progress, but society itself adapts and changes surprisingly little, much as it did when models passed the Turing test.

Culture pattern worth copying:

  • OpenAI’s edge, per Altman, is repeatedly doing new, unproven things rather than copying. Copying is easy once conviction exists; originating direction without that signal is what’s rare and what most orgs fail at.

Chapter Summaries

  • Future of OpenAI models: Continued investment in the o-series reasoning models is strategically central; expect rapid progress and major unlocks for science and code. No-code AI app builders for non-developers will come but not soon.
  • Where OpenAI will steamroll vs. where opportunity lies: Don’t build to patch current model weaknesses; build products and services that benefit as models improve. Trillions of dollars in new market cap will come from AI-native verticals like healthcare, education, and tutoring.
  • Value creation and capex: Massive capex will be matched by massive value creation; the exact multiple matters less than the direction. Healthcare and education alone represent trillions of dollars of upside.
  • Open source: Both open-source models and hosted APIs have a place; users will pick what fits.
  • Agents: Defined as long-duration, minimally-supervised task executors. The exciting frontier is massively parallel actions and “senior co-worker” agents, not just restaurant bookings. Pricing may shift to compute-based.
  • Models as depreciating assets and differentiation: True that they depreciate, but revenue justifies cost when you have a sticky product. Reasoning, multimodal, and new capabilities are OpenAI’s differentiation focus.
  • Research culture: OpenAI’s edge is doing new, unproven things rather than copying. Conviction unlocks replication, but originating direction is rare.
  • Leadership through hypergrowth: Hardest lesson is shifting from “next 10%” to “next 10x” thinking, including communication, planning, and long-lead infrastructure decisions.
  • Hiring young vs. experienced: Both work; the right rule is a high talent bar regardless of age. Young people bring fresh energy; experienced people are essential for high-stakes complex systems.
  • Coding models: Anthropic currently leads on coding; developers should use multiple models; framing will shift from “models” to “systems.”
  • Scaling laws: Altman believes scaling-law-driven improvement will continue “for a long time,” despite past failed runs and behaviors no one understood.
  • Decision-making under uncertainty: The hard part is the volume of 51/49 decisions; rely on 15-20 trusted domain advisors, not one generalist.
  • Semiconductors and complexity: Not the top worry but in the top 10%. The deepest risk is fractal coordination complexity across power, chips, networking, research, and product.
  • Analogies for AI: Internet and electricity are poor analogies; the transistor is best, with scaling laws and ubiquitous diffusion.
  • Quickfire: Build an AI-vertical product if starting today; book topic would be human potential; cursor team and the real-time API earn shout-outs; latency vs. accuracy should be user-controllable.
  • Leadership weakness: Altman currently feels product strategy is his weakest area; praises Kevin Weil’s discipline and user advocacy.
  • Five and ten-year vision: Breathtaking technological progress combined with surprisingly modest, healthy societal change.