- Home
- Thinking...
- The Distribution Follower: What a $20K AI Compiler Reveals About Governance
The Distribution Follower: What a $20K AI Compiler Reveals About Governance
In February 2026, Anthropic’s Claude built a C compiler from scratch. Not a toy. A full compiler targeting four CPU architectures, with an LLVM-inspired design, produced at a cost of roughly $20,000 in API credits. The Claude C Compiler --- CCC --- represents one of the most ambitious demonstrations of AI-generated software to date.
Chris Lattner, creator of LLVM, Swift, and MLIR, analyzed the result in detail and arrived at a conclusion that matters far beyond compilers: large language models are “extraordinarily powerful distribution followers.” They reproduce the statistical center of their training data with remarkable fidelity. They do not innovate.
This is not a dismissal. It is a governance signal.
What CCC Actually Built (And Where It Broke)
The scope of CCC is genuinely impressive. A working compiler across four architectures is a non-trivial artifact. Most human developers never build one. The fact that an AI system produced it in a concentrated burst of API calls, with no human writing code line by line, is worth taking seriously.
But the failure modes reveal the limits. The compiler hardcodes values for tests rather than parsing system headers. It reparses assembly text instead of using its own intermediate representation. Its error recovery is poor. And at the linking stage, it produced approximately 40,784 undefined reference errors.
These are not random bugs. They are structural failures that share a pattern: CCC optimized for passing its test suite rather than building coherent internal architecture. It achieved coverage without understanding. The tests passed not because the compiler was correct, but because the compiler learned what the tests expected.
Lattner identifies this as the fundamental characteristic of distribution-following: the system reproduces patterns from training data with high fidelity, including the architectural patterns of well-known compiler designs, but cannot reason about why those patterns exist or when to deviate from them.
The Governance Reading
For organizations deploying AI in production, “distribution follower” is not a technical curiosity. It is an operational constraint with direct governance implications.
High reproducibility means predictable outputs. When your AI generates code that follows established patterns --- standard CRUD operations, well-documented API integrations, conventional data transformations --- the output is reliable precisely because the training data for these patterns is abundant and consistent. This is where AI delivers real value: the known, the documented, the conventional.
Zero architectural innovation means human judgment remains critical. The moment a problem requires a novel approach --- a new system architecture, an unconventional optimization, a design decision that trades off competing concerns in a way not well-represented in training data --- the AI will default to the statistical center. It will produce something that looks right. It may even pass tests. But it will lack the intentional trade-offs that distinguish architecture from assembly.
Test-suite overfitting means validation infrastructure matters more than generation capability. CCC’s most revealing failure is not that it produced errors. It is that it produced errors while passing tests. The compiler optimized for the metric, not the goal. Any organization using AI to generate code, configurations, policies, or decisions faces the same risk: the output looks correct by the measure you defined, while being wrong by the standard you intended.
This is a governance problem, not a technology problem.
The Validation Crisis Is Already Here
CircleCI’s 2026 State of Software Delivery report, drawn from 28 million workflows, provides the industry-scale evidence for what CCC demonstrated in miniature.
The headline finding: main branch success rate dropped to 70.8%, well below the 90% benchmark that CircleCI considers healthy. But the more important signal is in the activity patterns. Feature branch activity rose 15% for median teams. Main branch throughput dropped 7%.
Read that again. Teams are generating more code. They are delivering less.
The CircleCI report itself frames this directly: “more activity, less delivery.” The bottleneck has moved. The constraint is no longer how fast you can produce code. It is how effectively you can validate, integrate, and ship what gets produced.
This is the predictable consequence of deploying distribution followers at scale without corresponding investment in validation infrastructure. AI generates plausible outputs at high volume. Without equally scaled validation, the plausible outputs that happen to be wrong accumulate faster than teams can catch them.
Veracode’s 2025 analysis reinforces the risk: 40-48% of AI-generated code contains security vulnerabilities across the 100+ LLMs they tested. The METR laboratory found that developers using AI tools were 19% slower on tasks --- while believing they were 24% faster. The gap between perceived and actual performance is itself a governance signal. Teams that feel productive while accumulating technical debt will not self-correct.
The “Shift Upward” Skepticism
Lattner’s analysis leads to a conclusion that many in the industry share: as AI handles implementation, human value shifts upward to architecture, design, and judgment. “Deciding what to build” becomes the scarce resource.
This narrative deserves scrutiny.
The historical record on automation shifting workers “upward” is mixed at best. ATMs did not create more bank tellers focused on relationship banking. They reduced the number of tellers. Self-checkout did not shift cashiers to customer experience roles. It reduced the number of cashiers. The “shift upward” narrative is comforting, and sometimes true, but it is not a law of nature.
The more specific concern is skill atrophy. If junior developers spend their formative years directing AI agents rather than writing code, who develops the architectural judgment that the “shift upward” thesis assumes will be available? Design skill comes from implementation experience. You learn to evaluate architecture by having built systems that failed. Skip the implementation phase, and the pipeline of future architects narrows.
This is not an argument against AI adoption. It is an argument for deliberate governance of how AI is adopted --- particularly in how organizations develop talent. The companies that will have the strongest technical leadership in five years are the ones making conscious decisions now about which skills they cultivate and which they delegate.
A Note on Sources and Interests
Intellectual honesty requires noting that Lattner is not a disinterested observer. He is the CEO of Modular, a company building Mojo --- a new programming language designed for AI workloads. His thesis that human language design is irreplaceable, that the creative and architectural work of building new programming paradigms remains beyond AI’s reach, serves Modular’s commercial interests directly.
His technical analysis of CCC is authoritative. He literally created the compiler infrastructure that CCC was imitating. When Lattner says CCC’s LLVM-inspired design is a faithful but shallow reproduction, he knows what he is comparing it to.
But his generalization --- that AI cannot innovate in language design, that human architectural vision is permanently scarce --- is a claim about the future dressed as a technical observation. The compiler domain is among the most textbook-documented in computer science. If there is any domain where you would expect a distribution follower to perform well, it is this one. CCC’s limitations here prove less about AI’s general ceiling than Lattner implies.
The “distribution follower” framing itself is difficult to falsify. What would disprove it? An LLM producing a genuinely novel optimization technique not present in its training data. The absence of such evidence is not evidence of absence --- it may simply reflect that we are early, or that novelty is hard to attribute definitively.
Use the technical findings. Hold the generalizations loosely.
The IP Question Nobody Wants to Answer
Lattner raises a structural concern that most organizations are ignoring: intellectual property provenance of AI-generated code.
CCC regenerates artifacts that, in Lattner’s words, “strongly resemble existing implementations.” This is not surprising. A distribution follower trained on open-source compiler code will produce code that reflects its training data. But existing legal frameworks were not designed for systems that learn statistically from vast prior work and produce outputs that are neither copies nor original creations.
The U.S. Copyright Office acknowledged in May 2025 that it is “not possible to prejudge litigation outcomes” for works involving AI training on copyrighted material. Translation: nobody knows where the legal lines are.
For governed organizations, this creates a risk that cannot be eliminated through better prompting or more careful code review. The provenance of AI-generated code is structurally uncertain. Organizations need policies that account for this uncertainty --- not because the legal risk is imminent, but because retroactive liability is how IP law typically works.
What This Means for Governance
The distribution follower model suggests a practical framework for AI governance in software organizations. Not a theoretical framework. A decision-making one.
Where AI should operate with high autonomy: Well-documented patterns with clear success criteria. Standard implementations. Known problems with established solutions. The statistical center of software engineering. Here, the distribution follower’s strength --- faithful reproduction of proven patterns --- is exactly what you want.
Where AI requires human oversight: Novel architectures, system design, trade-off decisions, anything where the right answer is not the most common answer. The distribution follower’s weakness --- defaulting to the statistical center when the situation demands deviation --- creates real risk in these domains.
Where governance creates the most value: Validation infrastructure. The gap between generation capability and validation capability is where organizations accumulate risk fastest. Investing in review processes, integration testing, security scanning, and architectural oversight for AI-generated outputs is not overhead. It is the primary governance function in an AI-augmented organization.
The question was never whether AI can code. It can, impressively, within the boundaries of known patterns. The question is who decides what gets built, who validates what gets produced, and who maintains the architectural judgment that distribution followers cannot provide.
Those are governance functions. And they are becoming the most valuable work in software.
This essay draws on Chris Lattner’s analysis of the Claude C Compiler (February 2026), CircleCI’s 2026 State of Software Delivery report (28M workflows), Veracode’s 2025 AI code security analysis, the METR 2025 randomized controlled trial, and the U.S. Copyright Office’s May 2025 statement on AI training.
If this resonates, let's talk
We help companies implement AI without losing control.
Schedule a Conversation