When Y Combinator's Garry Tan disclosed that approximately 25% of companies in the Winter 2025 cohort had codebases that were 95% or more AI-generated, the reaction split predictably between two camps. One read it as evidence that software engineering as a profession was being disrupted faster than anyone had anticipated. The other dismissed it as startup theater—impressive-sounding metrics masking shallow applications that would collapse under production load.
The more interesting analysis, which most commentary missed, is what the 95% aggregate figure actually describes about how the best early-stage software teams are working today and what it implies for the developers who come after them.
The 95% Figure in Context
The statistic needs unpacking before it can be interpreted. "AI-generated code" in the YC cohort context encompasses a wide range of contribution types: complete feature implementations written from a natural language specification, boilerplate scaffolding generated from framework templates, test suites generated from function signatures, documentation strings auto-completed during typing, and utility functions produced from a brief description.
Not all of these represent equivalent AI capability or equivalent engineer leverage. A codebase that is 95% AI-generated because a developer used GitHub Copilot for tab-completion tells a different story than one where an agent wrote complete API endpoints from a Notion spec while the founder focused on customer development.
The Winter 2025 cohort skewed toward the latter. What distinguished the high-AI-generation founders was not passivity—handing problems to a model and accepting whatever came back—but a particular kind of architectural discipline: they spent their human attention on the decisions that compounded, and delegated the work that did not.
What Founders Were Actually Doing
The working pattern that recurred across the most successful AI-heavy codebases in the cohort followed a recognizable structure.
| Founder Skill | What It Means | Why It Matters |
|---|---|---|
| Specification precision | Complete behavior specs with inputs, outputs, error conditions | Agents implement correctly without clarification |
| Architecture ownership | Human control over data models, API contracts, service boundaries | Hardest-to-reverse decisions stay with humans |
| Evaluation discipline | Reading agent output for edge cases, security issues, performance | Critical skill when 95% of code is AI-generated |
Specification precision over implementation attention. The founders getting the most leverage from AI code generation were not those with the best prompt engineering skills in a technical sense. They were the ones who had developed the discipline to specify desired behavior completely enough that an agent could implement it without requiring clarification. This turns out to be a deep software design skill dressed in new clothes: writing a complete specification for a function, with its expected inputs, outputs, error conditions, and performance requirements, is essentially the same cognitive work as writing a detailed function contract. The founders who had done this kind of disciplined specification work before—in test-driven development, in API design—picked up AI-directed development faster than those who had relied on exploratory implementation.
Architecture as the scarce resource. The most important engineering decisions in an early-stage codebase are the ones that are difficult to reverse: data models, API contracts, service boundaries, authentication architecture. These decisions are where AI-generated code is least reliable and where human judgment remains most valuable. The best founders in the cohort drew a clear line: AI writes the implementation, humans own the architecture. This is not a limitation of current models—it is a deliberate allocation of where expensive human attention creates the most durable value.
Evaluation as the primary engineering skill. When 95% of code is AI-generated, the critical skill shifts from writing code to evaluating it. Can you read a 200-line function produced by an agent and determine whether it correctly handles the edge case you care about? Can you spot the security vulnerability in the authentication middleware that the model generated? Can you identify the performance problem in the database query that looks syntactically correct but will cause issues at scale? These are different skills from implementation, and they are skills that experienced engineers have developed—which is why the cohort founders with engineering backgrounds outperformed those who treated AI as a substitute for engineering knowledge.
The Agentic Coding Stack That Made This Possible
The Winter 2025 cohort benefited from a generation of agentic coding tools that did not exist eighteen months earlier. The difference between AI code completion and agentic coding is the difference between a tool that predicts your next keystroke and a system that can execute a complete task cycle: read the existing codebase, understand the context, implement the requested change, run the tests, fix the failures, and produce a pull request.
The tools that enabled this—Cursor, Claude's coding capabilities, and agentic platforms that connect to development workflows—changed the unit of work from "line of code" to "feature." A developer who previously spent a day implementing a user authentication flow can now spend that day reviewing and refining an agent-generated implementation, then spending the reclaimed hours on product decisions.
This is the paradigm shift that Neumar is built around. Neumar's agentic coding capabilities, including its Linear ticket-to-PR pipeline, operationalize this working pattern for teams that want a structured workflow rather than an ad hoc prompt-and-review loop. The agent understands the ticket context, generates the implementation plan, executes against the codebase, and surfaces the result for review—compressing the "idea to code" cycle in ways that the YC cohort was discovering through their own experimentation.
The Counterargument: What the Metric Does Not Prove
The 95% figure deserves skepticism in one specific dimension: it does not say anything about codebase quality, correctness, or long-term maintainability. A codebase that is 95% AI-generated is not by definition better or worse than one that is 10% AI-generated. The metric measures origin, not quality.
The history of software development is littered with codebases that were generated rather than written—CASE tools in the 1980s, no-code platforms in the 2010s—that worked well in demo conditions and accumulated technical debt faster than handwritten equivalents. The current generation of AI-generated code is considerably more sophisticated, but the maintainability question is genuinely open. Code that is generated from a specification without deep understanding of the surrounding system has a tendency to produce locally correct but globally inconsistent solutions.
The founders who will build durable companies on AI-generated codebases are not the ones maximizing their generation percentage—they are the ones who maintain the architectural understanding needed to keep the generated code coherent over time. The 95% generation rate is impressive; the question is whether the remaining 5%—the human decisions—are concentrated in the right places.
Implications for Working Developers
If the YC Winter 2025 cohort represents a leading indicator of how software is going to be built at scale, the implications for working developers are significant but not as simple as "learn prompt engineering."
The developers who will remain most valuable are those who can do the things that AI generation currently does poorly: identify the failure modes in generated code, make architectural decisions that hold up over years rather than months, reason about system behavior under novel conditions, and communicate technical context to non-technical stakeholders in ways that improve decision quality.
These are all, notably, skills that require deep technical knowledge rather than less of it. The founders in the YC cohort who were building on the thinnest engineering foundations were the ones whose AI-generated codebases accumulated the most technical debt most quickly—not because AI wrote their code, but because they could not evaluate what the AI wrote.
The sustainable version of the 95% paradigm is not "anyone can build software now." It is "engineers who invest in deep understanding can build much more, much faster." That is a meaningful change, but it is different from the disruption narrative that the headline statistic tends to generate.
The best use of the time that AI coding tools return to developers is not more feature development. It is deeper architectural thinking, more rigorous evaluation, and more time understanding users—the inputs that determine whether the code being generated is the right code to write.
