The Force Multiplier They're Not Talking About
There’s a narrative gaining traction in every boardroom, every LinkedIn feed, every technology conference: AI is the great equalizer. A junior developer with the right tools can now output what a senior team used to produce. The playing field is leveling.
I’ve heard this argument. I’ve lived the counterargument.
Last August, I estimated a major platform initiative at 22 person-weeks of effort — 10 weeks, multiple developers, QA, business analysis support. Those were honest numbers, built from decades of scoping projects in a mature enterprise codebase.
Seven working days later, I shipped it. Alone, with an AI.
Not a prototype. A real-time rules engine evaluating thousands of securities every minute, 11 strategy types, an admin portal, 560 automated tests, zero regressions in a codebase with over 100 projects. And I added features that weren’t in the original plan.
The playing field is not leveling. If anything, the gap is widening — in favor of experience.
Here’s why.
Why is AI a force multiplier, not a capability replacement?
A force multiplier doesn’t replace what you bring. It amplifies it. Give a lever to someone with nowhere to stand and nothing happens. Give it to someone with thirty years of knowing exactly where to plant their feet, and you move the world.
The AI didn’t architect that rules engine. I did. The AI didn’t know that our microservices have 26 partitions providing natural parallelism, or that our caching layer has specific serialization quirks, or where a similar problem had already been solved three years ago in a different part of the codebase. I knew those things. The AI held them in context and executed against them with perfect consistency.
That’s the distinction that gets lost in the equalization narrative. AI fluency is not the differentiator. Everyone will have AI fluency shortly — it’s becoming a baseline, like knowing how to use a search engine. The differentiator is what you bring to the collaboration. And what you bring is your accumulated pattern library, your scar tissue, your judgment about what actually matters.
What does experience actually mean in the age of AI?
Most people think of experience as knowledge — facts accumulated over time. That’s the least interesting part of it.
Real experience — the kind that matters in high-stakes decisions — is compressed pattern recognition that operates faster than conscious thought. After three decades in technology, I carry the outcomes of thousands of decisions, hundreds of architectural choices, dozens of teams under pressure, multiple market cycles. None of that lives in a document. It lives as a felt sense that fires before I can articulate why.
When something feels wrong about an architectural direction before the analysis is complete — that’s not intuition in the mystical sense. That’s decades of pattern data surfacing faster than conscious reasoning can catch up. The research on expert decision-making is unambiguous: experts in complex, time-pressured domains don’t make their best decisions by reasoning harder. They make them by recognizing patterns and trusting the recognition.
AI doesn’t have that. AI has breadth — an extraordinary ability to survey the landscape of known solutions, generate options, stress-test logic, and hold enormous amounts of context simultaneously. What it cannot have is your specific scar tissue. The decisions you made that looked right and failed in production. The team dynamics that the org chart didn’t show. The client who said yes and meant maybe. The technology that won the benchmark and couldn’t survive contact with real data at scale.
Those patterns don’t exist in training data. They exist in you.
The blueprint is the proof
Earlier this year I published the process framework I’ve developed for AI-augmented development — a stage-gated pipeline that takes initiatives from idea to production. Six stages. Four of thinking, two of building.
That framework didn’t come from a methodology book or a conference talk. It came from thirty years of watching projects succeed and fail, distilled into a system and then applied to a fundamentally new paradigm.
Look at what it encodes. The stage-gating reflects the hard-won understanding that ungated projects collapse under their own momentum — a lesson you only learn by living through the collapse. The four-to-two ratio of thinking to building reflects decades of learning that the failure mode is almost always insufficient upfront clarity, not insufficient execution speed. The principles — KISS, YAGNI, incremental validation, write it down — each one is a scar with a lesson attached.
A technically proficient twenty-five year old with equal AI fluency could not have written that blueprint. Not because they lack intelligence or capability, but because they haven’t failed in the right ways yet. The blueprint is failure crystallized into foresight.
And then there’s this — the single most important line in the entire framework, easy to miss:
Stage 2 — Executive Review: This is the one stage where AI isn’t involved. Human judgment on strategic priorities.
Knowing where not to use AI is itself a function of experience. AI predicts patterns. It doesn’t carry fiduciary responsibility, understand internal politics, feel reputational damage, or know which risks are existential versus cosmetic. It produces possibilities. You decide. And the quality of that decision depends entirely on everything you’ve lived through to get here.
Speed is overrated. Judgment isn’t.
The 22 person-weeks to 7 days result is dramatic enough to make the point, but the number that actually matters more is this: zero regressions in a codebase with over 100 projects.
Speed without judgment creates chaos faster. The reason the result held — the reason it shipped clean into a mature enterprise environment without breaking anything — wasn’t AI capability. It was the judgment applied at every decision point along the way. Which patterns to follow. Which constraints were real and which were historical accidents. Which failure modes to design against. Where to test aggressively and where to trust the existing infrastructure.
Every one of those judgment calls drew on something the AI couldn’t provide. Every one of them reflected a pattern learned the hard way, somewhere in the previous thirty years.
What is the real danger of AI for experienced leaders?
If you’re an experienced leader watching younger colleagues ship AI-powered products at speed, the temptation is to feel behind. To focus on closing the technical literacy gap. To learn the tools.
Learn the tools. Absolutely. Technical fluency matters — not as the differentiator, but as the price of admission.
But don’t confuse fluency with leverage. The real danger isn’t that you’ll fall behind technically. The real danger — and this is the one worth worrying about — is outsourcing your judgment to the AI along with the execution.
AI is extraordinarily good at producing confident-sounding outputs. It will generate a strategy, an architecture, a roadmap, a recommendation — fluently, plausibly, instantly. And it will be wrong in ways that only someone with your experience can catch. The organizational constraint it doesn’t know about. The market dynamic it can’t fully model. The technical debt that will make that elegant solution a nightmare to implement in your specific codebase.
The moment you stop editing its outputs with your judgment — the moment you start accepting rather than directing — you’ve surrendered the only advantage that compounds over time.
Your job is not to become more like the AI. Your job is to become a better director of it.
The people who win
The people extracting the most from AI right now are not the most technically current. They’re not the youngest or the fastest or the most fluent in the latest models.
They’re the ones who hold both simultaneously. Deep pattern libraries from lived experience, and enough current literacy to know what the tools can and can’t do. They bring the judgment. The AI brings the bandwidth. Together, they produce results that neither could approach alone.
That’s a narrow band. It’s also exactly where three decades of building things puts you.
The playing field isn’t leveling. The leverage is finally catching up to the experience.
The throughline
This post is the third in an informal series. From 22 Person-Weeks to 7 Days was the result. The Technology Initiative Pipeline was the system. This is the explanation — why experienced people produce these results while others don’t, and why that gap will widen before it narrows.
The AI is the best tool I’ve ever used. It’s also only as good as the thirty years I bring to the collaboration.
That’s not a limitation. That’s the point.