On-demand exhaustive AI-analysis
Complete visibility into time & dollars spent
Create meaningful reports and dashboards
Track and forecast all deliverables
Create and share developer surveys
Align and track development costs
A diagram has been circulating in engineering circles lately, and it stopped me cold. It shows three stages of AI code adoption — Day 1, Day 100, and Day 1000 — with the ratio of AI-written code to human-written code growing over time.
But it's the captions underneath that tell the real story:
The original creator called these the "stages of grief." I think that framing is exactly right — but I'd go further. What we're witnessing is the accumulation of something I call comprehension debt.
Technical debt is a familiar concept: shortcuts taken during development that create future maintenance burden. You know you're accumulating it. You can see it in the codebase. You can plan to pay it down.
Comprehension debt is different. It's the growing gap between the code your team has shipped and the code your team actually understands.
And unlike technical debt, comprehension debt is often invisible — until something breaks and nobody knows why.
Here's what makes AI-generated code particularly dangerous: the code often works. The tests pass. The PR looks reasonable. But the engineer who approved it may not fully understand why it works, what assumptions it makes, or how it interacts with the rest of the system.
Multiply that across hundreds of PRs, dozens of engineers, and thousands of commits. That's comprehension debt compounding at a rate we've never seen before.
The diagram captures a progression that I'm seeing play out in real engineering organizations:
AI-generated code is a productivity supercharger. Engineers review PRs with full context. The codebase is still mostly human-written, so the mental model holds. Velocity increases. Leadership celebrates.
What it feels like: "This is amazing. We're shipping so much faster."
The hidden risk: Teams start optimizing for speed, not comprehension. Review rigor quietly slips.
The ratio shifts. More code is AI-generated than human-written. Engineers start reviewing PRs where they understand the intent but not every implementation detail. The phrase "it probably won't break anything" enters the vocabulary.
What it feels like: "I trust the AI. It's been reliable so far."
The hidden risk: Institutional knowledge erodes. New engineers onboard to a codebase that even senior engineers can't fully explain. The cognitive gap widens.
The codebase has grown beyond any single engineer's comprehension. PRs get approved based on test results, not understanding. When something breaks, debugging becomes archaeology. The "YOLO" merge becomes normalized.
What it feels like: "I don't really know how this works anymore, but it seems fine."
The hidden risk: You're one complex incident away from a crisis nobody on the team can resolve without AI assistance — and AI doesn't take accountability for production outages.
Individual engineers aren't going to solve this on their own. The incentives are misaligned: shipping fast gets rewarded, slowing down to understand doesn't. Comprehension debt is a team-level and organizational-level risk that requires leadership intervention.
Here's the uncomfortable truth: most engineering organizations are somewhere between Stage 1 and Stage 2 right now. They're in the honeymoon phase — or just starting to drift. The cliff is still ahead of them. But at the rate AI code generation is accelerating, that cliff is closer than they think.
The question isn't whether your team will accumulate comprehension debt. It's whether you're measuring it, managing it, and building the systems to stay ahead of it.
If you're an engineering leader, here are the signals to watch for:
The obvious question: if AI is generating the code, can't AI also review it and close the comprehension gap?
I've been tracking this space closely, and the honest answer is: not yet.
A recent discussion on Hacker News captured the current state well. Developers acknowledge that AI-powered code review tools can identify critical bugs that humans might overlook. That's real value. But the primary drawback is a poor signal-to-noise ratio — excessive, speculative comments that create more work, not less.
The consensus among practitioners is telling: AI code review currently functions as an advanced linter, but lacks the contextual and architectural understanding of a human reviewer. It can catch syntax issues and common patterns. It struggles with "does this actually make sense for our system?"
Some firms like Jane Street have found alternative workflows — automated style enforcement combined with editor-integrated reviews — that extract value without the noise. And many developers are skipping the specialized "wrapper" tools entirely, preferring to use frontier models directly when they need AI assistance.
The takeaway isn't that AI code review is useless. It's that AI tooling alone cannot replace the nuanced judgment required for high-level software design. Which means the comprehension problem still lands on humans — and on the organizational structures we build to support them.

(Source: https://news.ycombinator.com/item?id=46766961)
Comprehension debt isn't going away. AI code generation is accelerating, not slowing down. The answer isn't to reject these tools — it's to build the organizational muscle to stay ahead of the debt.
Add a specific checkbox or prompt: "Can you explain how this code works without referencing the AI prompt?" If the answer is no, that's a flag — not necessarily a blocker, but a signal that comprehension debt is being incurred.
People are already experimenting with interesting ways to solve this. The most feasible approach I've seen comes from Geoffrey Litt, who proposed popping a short quiz to code reviewers — a few questions about the code they're approving to verify whether they actually understand it. It's a lightweight friction that forces the reviewer to pause and confirm comprehension before hitting "approve."
This isn't about slowing teams down. It's about creating a moment of intentionality. The quiz doesn't need to be exhaustive — even two or three questions about control flow, edge cases, or integration points can surface whether the reviewer has a mental model or is just rubber-stamping.
Know what percentage of your codebase is AI-generated. Know which modules have the highest concentration. Know which engineers are approving the most AI code. This isn't about surveillance — it's about visibility into where comprehension debt is likely accumulating.
Observability, tracing, and repository intelligence tools become critical when human comprehension lags behind code volume. If your engineers can't explain the code, your systems need to — through detailed logging, contextual debugging, and AI-assisted code exploration.
Incident response, on-call rotations, and debugging sessions force engineers to understand systems deeply. Protect this time. Don't let AI-assisted velocity crowd out the work that builds real understanding.
Velocity metrics (commits, PRs merged, features shipped) will all look great as comprehension debt accumulates. Add countervailing metrics: incident resolution time, onboarding ramp time, code review depth, and architecture decision documentation. If velocity is up but these are degrading, you're borrowing against the future.
Here's the thing that keeps me up at night as an investor and advisor: the computer can never be held accountable.
When AI generates code and something catastrophic happens in production, the engineer who approved that code is still on the hook. "Claude wrote it" is not a defense. "The tests passed" is not a defense. The human in the loop remains responsible.
Comprehension debt undermines that accountability. You can't be responsible for code you don't understand. But you're responsible anyway.
That's the bind engineering leaders are in right now. And it's why treating comprehension debt as a first-class organizational risk — not just an individual engineer problem — is so critical.
I remain deeply optimistic about AI-native software development. Jevons paradox tells us that efficiency gains increase demand — and I believe we'll have more software engineers, not fewer, creating more value than ever before.
But that optimistic future only arrives if we build the organizational muscle to manage what AI acceleration produces. That means:
Most engineering teams are still in the green zone. Some are drifting into yellow. The goal is to never hit red — or at least to see it coming and course-correct before the YOLO merge becomes your only option.
The code generation revolution is real. The comprehension debt crisis is coming. The engineering leaders who see it early will be the ones still standing when things break.

Diagram showing the three stages of comprehension debt: Day 1 (Honeymoon), Day 100 (Drift), Day 1000 (Cliff) — with ratio of AI-written to human-written code growing over time.
Alok Nandan is a General Partner at First Ray Ventures, a seed-stage fund investing in developer tools, DevOps, and infrastructure companies. He writes about the evolving role of engineering leadership in the AI era. Connect with him on LinkedIn.