Andrej Karpathy’s Vision of Software 3.0 at YC AI Startup School

image

[image found/credit inc.com]

In June 2025, Andrej Karpathy, a renowned AI scientist known for his work at Stanford, OpenAI, and Tesla, delivered a keynote at Y Combinator’s inaugural AI Startup School in San Francisco. His talk, titled “Software in the Era of AI,” introduced the concept of “Software 3.0,” where large language models (LLMs) and natural language prompts redefine how software is built. Speaking to 2,500 AI students and founders, Karpathy outlined a future where AI augments human creativity, not replaces it, and urged startups to embrace this shift thoughtfully. His insights sparked excitement, with many calling it a defining moment for understanding AI’s role in software development. Below is a summary of his talk, followed by key points from the event.

Karpathy began by tracing software’s evolution. Software 1.0 was traditional coding in languages like C++ or Python, where humans wrote explicit instructions. Software 2.0, which he pioneered at Tesla, used neural networks to replace hand-written code with trained models, like those for autopilot systems. Now, Software 3.0 marks a new era where LLMs, programmed through English prompts, act as a new kind of computer. He called English “the hottest new programming language,” noting that anyone can now create software by describing what they want, a process he termed “vibe coding.” This democratizes coding, letting non-technical founders build apps quickly.

However, Karpathy cautioned that LLMs are not perfect. He described them as “people spirits”—simulations of human intelligence trained on internet data, with superhuman strengths like encyclopedic knowledge but also flaws like “jagged intelligence.” For example, LLMs can solve complex math problems but might claim 9.11 is larger than 9.9 or miscount letters in “strawberry.” They also suffer from “anterograde amnesia,” lacking long-term memory beyond their context window, and are prone to hallucinations and security risks like prompt injection. To address this, he advised keeping “AI on a leash” with human oversight.

Karpathy predicted the next decade (2025–2035) will be the “decade of agents,” but not fully autonomous ones. Instead, he championed “partial autonomy” apps like Cursor and Perplexity, which blend AI with human verification. These tools use custom interfaces, manage context, and offer “autonomy sliders” to adjust AI control, from simple code completion to file-level changes. He emphasized that successful AI apps need fast, visual verification loops to ensure humans can audit AI outputs easily, avoiding errors in critical tasks like coding or research.

He also compared LLMs to utilities, semiconductor fabs, and 1960s mainframe operating systems. Like utilities, they require huge upfront costs to train and ongoing costs to serve via APIs. Like fabs, they demand deep R&D investment. Like mainframes, they run in the cloud, sharing compute across users, with chat interfaces acting as modern terminals. Unlike traditional tech, LLMs spread from consumers to enterprises, flipping the usual diffusion pattern. This shift, he said, opens massive opportunities for startups to build AI-native tools and infrastructure.

Karpathy urged founders to focus on practical applications, not overhyped AGI dreams. He suggested starting with vibe coding for simple projects but warned that scaling startups need skilled developers for robust systems. He also called for new standards, like “llms.txt” files, to make software LLM-friendly, easing agent interactions. His talk ended with a call to action: embrace Software 3.0’s potential while staying grounded in human-AI collaboration to build reliable, innovative products.

Key Points from Andrej Karpathy’s Talk at YC AI Startup School

1. Software 3.0 Defined: Software has evolved from traditional code (1.0) to neural networks (2.0) to LLMs programmed via natural language prompts (3.0). English is now a programming language, enabling “vibe coding” where anyone can build software by describing ideas.

2. LLMs as New Computers: LLMs act like cloud-based operating systems, with models as CPUs, context windows as memory, and prompts as code. They resemble 1960s mainframes, with shared compute and high costs, but diffuse from consumers to enterprises.

3. LLM Strengths and Flaws: LLMs are “people spirits” with superhuman knowledge but suffer from jagged intelligence (e.g., failing simple comparisons like 9.11 vs. 9.9), anterograde amnesia (no long-term memory), hallucinations, and security risks.

4. Partial Autonomy Apps: The future lies in apps like Cursor and Perplexity, which combine AI with human verification, custom GUIs, and autonomy sliders. Fast, visual verification loops are key to keeping “AI on a leash” and avoiding errors.

5. Decade of Agents (2025–2035): Karpathy predicts AI agents will augment, not replace, humans. Full autonomy is unreliable; instead, focus on tools that enhance productivity, like coding assistants or research aids.

6. Startup Opportunities: Founders should build LLM-friendly infrastructure (e.g., llms.txt, APIs) and practical apps, not chase AGI. Vibe coding is great for prototyping, but scaling requires skilled developers for robust systems.

In conclusion, Karpathy’s talk was a roadmap for navigating the AI-driven future of software. By framing LLMs as both powerful and fallible, he inspired founders to build tools that amplify human creativity while addressing AI’s limits. His vision of Software 3.0 offers a practical yet exciting path for the next generation of AI startups.

Credit: Andrej Karpathy.