The Agentic SDLC: Uniting the AI Tool Sprawl
Every tool in your product development life cycle is now an AI agent trying to do everything. Here is how to stop the chaos, draw the right boundaries, and build an orchestrated pipeline that actually works.
The Agentic SDLC: Uniting the AI Tool Sprawl
There is a problem hiding inside the AI tooling revolution, and most engineering organizations are already living it. We have never had more AI capability across the product development life cycle. Jira uses AI to generate stories. Figma turns wireframes into React components. GitHub Copilot writes the code, Claude Code rewrites it, and Cursor refactors it. The power is extraordinary. The clarity is gone.
Picture this: a product manager uses Jira's AI to generate a user story and spin up a component spec. A designer uses Figma's AI to turn a wireframe into React code. An engineer feeds that same wireframe into Claude Code and gets a different component entirely. Three AI tools, three outputs, zero shared context, and nobody on the team realizes it until someone tries to integrate them on Friday afternoon. This is not an edge case. It is the daily reality in organizations that added AI tools faster than they added a framework to use them. I have watched this play out firsthand, and the frustration it creates is real.
The result is a sprawling, redundant, and increasingly incoherent landscape of AI tools spinning in circles. We need to stop thinking about AI tools and start thinking about AI orchestration. Not what each tool can do, but what role each one plays, what artifact it produces, and how that artifact moves downstream.
Here is the uncomfortable truth underneath all of this: AI did not create the broken handoffs between product, design, and engineering. It exposed them. The dysfunction was already there. We just never shipped fast enough for it to hurt. Now we do. And now it does.
What we are really building, whether we intend to or not, is an agentic software development life cycle. The question is whether we build it deliberately or inherit it by accident.
"The bottleneck has inverted. Writing code is no longer what slows you down. Knowing what to build, and communicating it clearly, is the last expensive human problem."
The Hourglass Inversion
For decades, the software engineering phase was the longest, most expensive part of the product development life cycle. The pipeline was a funnel that narrowed to a bottleneck every time work hit implementation.
AI has flipped this completely. Today, a skilled engineer can produce in hours what used to take weeks. For well-defined, well-specified problems, the code generation problem is essentially solved. Novel architecture, complex distributed systems, and deep domain logic still require experienced engineering judgment. But for the majority of standard implementation work, the bottleneck is no longer writing the code. This creates a new and uncomfortable truth: the constraint has moved upstream.
I see this in my own teams. The investment in AI tools at the code layer pays off only when the layers above it produce clear, structured, machine-consumable work. If your team has not yet reckoned with this inversion, you will simply ship "plausible-looking garbage" faster. That is not a win.
Why Every Tool Is Trying to Do Everything
Before going further, I want to be precise about language. Most of what we call "AI agents" in these tools are not autonomous agents in any meaningful technical sense. Jira's AI generates text. Figma's AI exports code patterns. Copilot completes your current line. These are AI-assisted features: powerful, context-limited, and operating entirely inside their vendor's sandbox. True agentic systems plan, use tools, and adapt across multiple turns toward a goal. That distinction matters when you are designing a pipeline, because the failure modes are completely different.
With that grounding, the boundary problem is real and it is structural. Every AI tool vendor is operating under the same competitive pressure: expand the surface area before a competitor does. Jira does not want to be a ticket system. It wants to be the intelligence layer for your entire delivery operation. Figma does not want to export designs. It wants to be the bridge between intent and code. Each tool is racing toward the same north star: end-to-end ownership of the product delivery pipeline.
This is not malicious. It is rational market behavior. But it creates a genuine problem for the organizations consuming these tools. When every tool claims to cover the full life cycle, you end up with duplicated context, inconsistent outputs, and AI features optimized for their own domain making decisions in domains they were never designed for.
A Jira AI that generates code will produce code that fits neatly into a Jira ticket structure. It will not generate code that understands your architectural patterns, your security posture, or your deployment constraints. A Figma AI that writes components will produce components that render correctly. It will not produce components that conform to your design system, pass your accessibility standards, or integrate cleanly with your existing state management. The tool's context is narrow. The tool's confidence is high. That is a dangerous combination.
The solution is not to pick one tool that does everything. No such tool exists, and the ones that claim to are lying. The solution is to define what each layer of the life cycle is responsible for producing and to make that output the unit of handoff between layers.
A New Mental Model: The Intent Stack
The most useful reframe is this: stop organizing around tools and start organizing around layers of intent. AI can move effectively within a layer, generating variations, filling gaps. It is unreliable at crossing layers without structured input.
The Agentic PDLC / SDLC: Intent Stack Model
| Layer | Responsibility | Primary Output | Owner |
|---|---|---|---|
| 1. Vision & Strategy | Business goals, OKRs, user problems | Strategic Brief | Product Leadership |
| 2. Requirements | User stories, API contracts, data models | The Context Spine | Product + Engineering |
| 3. Design | Interaction models, component specs | Annotated Design | Design + Engineering |
| 4. Implementation | Code generation, unit testing, security scanning | Reviewed & Tested Code | Engineering (AI-Agentic) |
| 5. Validation | Integration, performance, security gates | Verification Report | Engineering (AI-Agentic) |
| 6. Feedback | User behavior, outcome measurement | Insight Loop | Product + Engineering |
One thing this table makes explicit that most engineering teams miss: Layer 2 belongs to Product. Layer 3 belongs to Design. This framework does not work if engineering is writing requirements because the PM handed over a slide deck. The Context Spine requires a genuine partnership across all three disciplines at the top of the stack. If your product-to-engineering handoff is broken today, accelerating the implementation layer will not fix it. It will surface the breakage faster and more expensively.
The Context Spine: The Connective Tissue
The framework above only works if there is a structured artifact at each layer transition. Something the next layer can actually consume without guessing. This is what I call the Context Spine: a living specification document that every layer reads from and writes to, and that travels with the work through the entire pipeline.
Most organizations do not have this today. They have a Jira ticket with three lines of description and a Figma file with no annotations. When the implementation layer tries to execute based on that input, it is guessing. AI execution on a vague spec is not slower than human execution on a vague spec. It is faster and more confidently wrong.
The Context Spine changes this. It is not a heavyweight document. In practice, five fields cover most situations: a one-paragraph problem statement, the acceptance criteria, a list of explicit constraints (security, compliance, performance boundaries), the expected interfaces (inputs and outputs), and a list of non-goals. That is it. Machine-readable. Repo-resident. Version-controlled alongside the code it governs.
There is also a technical reason the Context Spine matters beyond process clarity. Language models are stochastic. The same input can produce meaningfully different outputs across runs. That non-determinism is manageable, but it requires your spec to be precise enough that variance in the output is easy to detect and reject. In a regulated environment, "close enough" is not an engineering standard. The Context Spine is part of how you make that precision architectural rather than aspirational.
Leadership Metric: Do not just measure "Developer Velocity." Measure Spec-to-Code Alignment. If your AI output is deviating from intent, your bottleneck is a lack of clarity in your requirements, not a lack of speed in your IDE. One practical proxy: after each sprint, have engineers flag the percentage of generated code that required structural rework, not style cleanup, but changes to logic, data models, or integration contracts. Even a rough weekly number reveals where your specs are leaking before the pattern becomes a liability.
The teams I lead have been building toward this with a pipeline-first delivery model. No direct production access. Every change moves through a defined pipeline with defined gates. The same principle applies here: no output bypasses the Context Spine. No layer gets skipped. Every transition is an opportunity for human review — not a mandate for it, but an opportunity the pipeline makes easy to exercise when something looks wrong.
Six Principles for Getting This Right
01. Spec Before Speed
Resist the temptation to generate code before the spec is solid. Poor specs now cost more than ever because they trigger high-speed, incorrect execution. A vague input to a fast agent is not a productivity gain. It is a fast path to rework.
02. Tools Serve Layers, Not the Reverse
Choose tools based on the layer they serve, and be explicit about the boundary. If your team is using Cursor to generate requirements or prompting Jira's AI to write implementation code, the problem is not the tools. It is the absence of a layer model. Assign each tool to a layer and hold that line.
03. Produce Consumable Artifacts
Every layer must produce a structured output — Markdown, JSON, OpenAPI — that the next layer can consume without "interpreting vibes." If the downstream consumer has to infer intent, the artifact is not done.
04. Own the Context Spine
The spec document that lives in your repo is more valuable than any individual tool. Treat it as the source of truth for every agent in the pipeline. When the spec drifts, everything downstream drifts with it.
05. Checkpoint Transitions, Not Operations
Human review belongs at layer transitions, not inside every operation. This is not about removing accountability. It is about placing review at the right altitude. Engineers reviewing every line of AI-generated code are not leveraging the technology. Engineers reviewing whether the implementation matches the spec are doing exactly the right thing.
06. Close the Feedback Loop
Shipping fast is easy now. Learning fast from what you shipped is the differentiator. Layer 6 is the least built-out layer in most organizations and the most strategically valuable. A production signal that never makes it back to Layer 1 means you are optimizing for output, not outcomes. Build the feedback loop deliberately. Instrument what users actually do, not just what they say. Connect that signal to your requirements layer so the next spec starts with evidence, not assumptions. Without this, you are not building an agentic SDLC. You are building a faster waterfall.
The Path Forward
The organizations that win over the next five years will not be the ones with the most AI tools. They will be the ones that built a coherent, orchestrated pipeline where every layer has a clear owner, every agent has a clear role, and humans are positioned at the decisions that actually require judgment.
Here is where I would start. Audit your current tool landscape against the Intent Stack. Identify where context is lost between layers. That is your highest-value first investment. Then instrument one complete path from requirements to deployment with a Context Spine, measure Spec-to-Code Alignment, and iterate.
One honest caveat: this is organizational change work, not just tooling work. Vendor contracts exist. Engineers have preferences built over years of practice, and some of that resistance is not irrational. It is identity, craft, and legitimate skepticism about AI output quality. You are not going to reorganize your entire pipeline in a sprint, and you should not try. Pick one team, one product area, one complete path through the stack. Make it work. Then let the results do the arguing for you.
The agentic SDLC is not a product you buy. It is a discipline you build. The tools exist. The capability is real. What most organizations are missing is the architecture that makes it all add up.
"The agentic SDLC is not a product you buy. It is a discipline you build."
The bottleneck has moved. Move with it.
Share this post

Written by Vinny Carpenter
VP Engineering · 30+ Team Experience
I lead engineering teams building cloud-native platforms at a Fortune 100 company. I write about engineering leadership, AI-assisted development, platform strategy, and the hard lessons that come from shipping at scale.
Related Posts
Eleven Years, One Week, and an AI Co-Pilot: Rebuilding TravelTimes for iOS 26.4
A decade-old side project, six major features, one week. How spec-driven AI-assisted development compressed months of work into a focused sprint on a real codebase with real constraints — and where the AI got it wrong.
How I Built a Production Notification System for TravelTimes in One Day with Claude Code
From 'users want commute alerts' to 1,800 lines of shipped, App Store-ready code in a single coding session. A deep dive into architecture, edge cases, and what AI-assisted iOS development actually looks like.
The New Abstraction Layer: AI, Agents, and the Leverage Moving Up the Stack
Andrej Karpathy put words to something many engineers are quietly feeling right now. We've been handed a powerful alien tool with no manual, and we're all learning how to use it in real time.