Ch 7 — AI-Native Teams & the New SDLC

How teams restructure when agents handle the repeatable work
High Level
lightbulb
Design
arrow_forward
description
Spec
arrow_forward
smart_toy
Build
arrow_forward
rate_review
Review
arrow_forward
rocket_launch
Ship
arrow_forward
monitoring
Observe
-
Click play or press Space to begin...
Step- / 8
groups
The Fundamental Shift
People own outcomes, agents handle repeatable work
What Changes
In a traditional team, developers spend roughly 60–70% of their time on implementation — writing code, fixing bugs, updating tests, handling migrations. In an AI-native team, agents handle much of that implementation work. Developers shift to design, specification, review, and exception handling. The ratio flips: 30% implementation, 70% design and review. This isn’t a reduction in work — it’s a change in the type of work.
What Doesn’t Change
Humans still own the outcomes. The agent can write the code, but a human decides what to build, how to architect it, whether the implementation is correct, and when to ship. The accountability chain doesn’t change — if the agent introduces a bug, the human who approved the PR is responsible. This is why review skills become the most valuable skill in the new model.
Key insight: AI-native doesn’t mean “fewer developers.” It means developers doing different, higher-leverage work. The team that used to ship 10 features a quarter now ships 30 — with the same headcount.
person
New Roles Emerging
The jobs that didn’t exist two years ago
Agent Operator
Manages the agent fleet: writes task descriptions, monitors agent performance, tunes blueprints, and handles escalations when agents fail. This is a hybrid role — part developer (understands the codebase), part ops (manages the system), part product manager (prioritizes the task queue). In smaller teams, this is a responsibility, not a full-time role.
Review Specialist
As agent output scales, review becomes a bottleneck and a specialization. Review specialists focus on evaluating agent-generated PRs: checking for correctness, architectural fit, security implications, and edge cases. They develop expertise in recognizing common agent failure patterns and can review faster because they know what to look for.
Key insight: These roles emerge naturally as teams scale agent usage. You don’t need to hire for them on day one — they evolve from existing developer roles as the team’s relationship with agents matures.
description
The Spec Becomes the Product
When agents build from specs, spec quality determines output quality
Why Specs Matter More
In the old model, a vague ticket was fine because the developer would ask clarifying questions during implementation. In the agent model, the spec is the input to the agent. A vague spec produces vague code. A precise spec with clear acceptance criteria, edge cases, and constraints produces precise code. Writing good specs becomes a core engineering skill — as important as writing good code used to be.
What a Good Agent Spec Looks Like
Context: what exists today and why it needs to change. Requirements: specific, testable acceptance criteria. Constraints: what not to touch, patterns to follow, performance requirements. Test strategy: how to verify the implementation. Scope boundary: explicit statement of what’s out of scope. The more precise the spec, the better the agent output.
Key insight: The irony: as coding becomes more automated, the ability to clearly articulate what you want in natural language becomes the differentiating skill. The best “AI-native developers” are excellent writers.
conversion_path
How Every SDLC Stage Changes
The new software development lifecycle
Planning & Design
Before: Architect designs, team estimates, PM prioritizes. After: Same, but estimates change dramatically. Tasks that took a developer 2 days now take an agent 2 hours + 30 minutes of review. This changes what’s feasible in a sprint. The backlog of “nice to have” items becomes achievable.
Implementation
Before: Developer writes code. After: Developer writes spec, agent writes code, developer reviews. For complex features, the developer still writes the architecture and critical paths; agents handle the scaffolding, tests, and boilerplate.
Testing & QA
Before: Developers write tests, QA tests manually. After: Agents generate test coverage, maintain the suite, and catch regressions. QA focuses on exploratory testing, UX validation, and edge cases that require human judgment.
Deployment & Operations
Before: CI/CD runs, humans monitor. After: AI agents in the pipeline catch issues pre-merge. Post-deploy, agents can detect anomalies and propose rollback PRs. Humans make the final call on rollbacks.
speed
New Metrics That Matter
Lines of code is dead. Here’s what replaces it.
Output Metrics
Features shipped per sprint — the ultimate measure of team throughput. Time from spec to merged PR — how fast can the team go from idea to production? Agent leverage ratio — how many agent-hours does each developer-hour produce? A developer who spends 1 hour writing specs that generate 8 hours of agent work has an 8x leverage ratio.
Quality Metrics
Review acceptance rate — percentage of agent PRs merged without major rework. Production incident rate — are agent-generated changes causing more or fewer incidents? Technical debt velocity — is the team paying down debt faster with agent help? Developer satisfaction — are developers happier doing design and review than writing boilerplate?
Key insight: The most dangerous metric is “PRs merged per day.” Without quality gates, teams can game this by merging low-quality agent output. Always pair throughput metrics with quality metrics.
psychology
What Humans Still Do Better
The irreplaceable human skills
Judgment Under Ambiguity
When requirements are unclear, when there are multiple valid approaches, when the trade-offs involve business context the agent doesn’t have — these are human decisions. “Should we optimize for speed or maintainability?” requires understanding the team’s roadmap, the product’s lifecycle stage, and the company’s priorities. No agent has this context.
System Design & Architecture
Agents are excellent at implementing within a design; they’re poor at creating the design. Choosing between a monolith and microservices, deciding on a data model, designing an API contract — these require understanding the full system, the team’s capabilities, and the long-term evolution of the product. Architecture remains a deeply human skill.
Key insight: The skills that become most valuable in an AI-native team are the ones that were always valuable but underappreciated: system design, clear communication, code review, and the judgment to know when the agent is wrong.
warning
The Transition Pitfalls
Common mistakes teams make when adopting autonomous workflows
Pitfall 1: Deskilling
If junior developers never write code from scratch, they never develop the deep understanding needed to review agent output effectively. The solution: deliberate practice. Junior developers should still write critical code manually, with agents handling the routine work. The goal is to accelerate learning, not bypass it.
Pitfall 2: Review Fatigue
When agents produce 20 PRs a day, reviewers get overwhelmed and start rubber-stamping. The solution: scale review capacity with agent output. Use AI-assisted review tools. Set a maximum PRs-per-reviewer-per-day limit.
Pitfall 3: Spec Debt
Teams rush to submit tasks to agents without writing good specs, then waste time fixing bad output. The solution: treat spec writing as a first-class engineering activity. Review specs before submitting them to agents, just as you’d review a design doc.
Key insight: Every pitfall has the same root cause: treating agents as a shortcut instead of a tool that requires new skills. The teams that succeed invest in the human skills (spec writing, review, architecture) alongside the agent infrastructure.
rocket_launch
Building an AI-Native Culture
It’s a culture change, not just a tool change
Start with Champions
Don’t mandate agent adoption top-down. Find 2–3 developers who are excited about the technology and let them experiment. Let them share their successes (and failures) with the team. Organic adoption driven by visible results is more sustainable than mandated adoption driven by management enthusiasm.
Celebrate the Right Things
Celebrate features shipped, not code written. Celebrate thorough reviews that caught agent mistakes, not just fast merges. Celebrate well-written specs that produced great agent output. The behaviors you celebrate are the behaviors that scale. If you celebrate “most PRs merged,” you’ll get quantity over quality.
Key insight: The transition to AI-native teams is a 6–12 month journey, not a switch you flip. It requires new skills, new metrics, new review processes, and a cultural shift in what “good engineering” means. Be patient with the transition.