Contents
The “One Engineer Can Build Anything” Temptation
That shift is why the recurring “one engineer can now build an application” narrative is both directionally true and strategically incomplete.
There are contexts in which a single engineer, working with strong tools, can do something today that would have taken a small team not long ago. That is real. It is also most visible in bounded settings: internal tools, prototypes, greenfield applications, and work where implementation is the dominant constraint. But most real engineering organizations are not operating in those conditions all the time. They are dealing with existing systems, multiple services, legacy assumptions, operational obligations, security constraints, and the accumulated complexity of software that has been alive long enough to matter. In those conditions, the bottleneck is not simply code production. It is coordination, correctness, and continuity.
The emerging academic evidence points in the same direction. A recent Science paper on the global diffusion of AI coding assistants found measurable productivity benefits associated with AI-assisted coding, but those benefits were concentrated among senior developers, while early-career developers showed no statistically significant gains. Another recent randomized study of experienced open-source developers working in mature repositories found something even more useful as a caution: in that setting, AI actually increased completion time by 19%. Neither of these studies invalidates the broader productivity story. What they do is force the more interesting conclusion: gains are real, but they are heterogeneous. They depend on codebase maturity, workflow design, user experience, and the degree to which the person using the tool has the context and judgment to direct it effectively.
That is one reason I do not think the right executive response is to push engineering teams toward a simplistic mandate to “use AI more.” A better response is to standardize the conditions under which AI actually produces leverage.
Two Modes: Inline Help and Task Execution
At Liferaft, that means distinguishing between two different modes of use. One is inline assistance: quick help while coding, small implementations, localized refactors, scaffolding, and test generation. The other is broader task-oriented execution: multi-file changes, ambiguous work, larger features, migrations, non-obvious debugging, and planning. In practical terms, that is why I think the model of Copilot in the IDE and Codex CLI-first makes sense. The point is not that these are the only tools that matter. The point is that they reflect two very different kinds of engineering work, and that work benefits from different modes of interaction.
What matters even more than the tools, though, is the workflow around them.
For small tasks, the engineering loop can remain lightweight. Use AI inline, move quickly, generate or extend tests, ask for a quick self-review, then submit. For larger or less bounded work, the workflow should be more deliberate: begin with analysis, ask for a plan before asking for code, review that plan, generate a first pass, refine locally, expand tests, run checks, inspect the diff, and then ask AI to critique the change before it reaches another human reviewer. This is not process for its own sake. It is how you convert an assistant from a faster autocomplete engine into something closer to a productivity multiplier.
That distinction is important because there is now enough evidence to show that adoption alone does not produce durable gains. DORA’s AI Capabilities Model makes that point explicitly: organizations realize more value when AI adoption is accompanied by stronger systems, clearer workflows, and more mature engineering practices. In other words, AI does not rescue weak engineering discipline. It rewards strong engineering discipline.
This is also why I do not think AI in engineering should be framed defensively — either by leaders or by engineers themselves.
The wrong way to talk about it is as if the craft is being hollowed out, or as if the core skill of the profession is about to disappear. The more accurate interpretation is that some forms of labor are becoming less scarce, while some forms of judgment are becoming more valuable. When a draft can be generated quickly, the quality of the outcome depends even more on the quality of the person steering it. Architecture matters more. Review matters more. The ability to reason about failure modes matters more. Taste matters more. Knowing when not to accept the obvious solution matters more.
If that sounds familiar, it should. Most technological shifts do not erase expertise. They move its center of gravity.
Pillar 1 as the Beginning, Not the End
That is what I think Pillar 1 really is.
It is not the end state. It is the beginning of a broader change in how engineering work is organized. But it matters because it is where the organization learns the first real lessons: how to distinguish speed from throughput, how to redesign workflows rather than just adding tools, how to preserve accountability while increasing leverage, and how to make stronger engineers more effective rather than merely making output more abundant.
That is also why I think the right expectation is not magic.
The best available evidence supports real gains, but not fantasy. In most organizations, a well-executed Pillar 1 should be thought of as a meaningful productivity and quality initiative, not as a promise that one person now replaces many. The real payoff is that engineering teams can spend less time on low-leverage work, move faster through routine implementation, strengthen tests and self-review, and apply more of their attention to the parts of software development that were always hardest to automate in the first place.
That is already significant.
And it is only the first pillar.