DORA defined an era of engineering performance. Its four metrics1 gave engineering leaders a shared vocabulary:
- Lead Time for ChangesLead Time for ChangesElapsed time from a code change being committed to that change running in production. Measures the deployment pipeline.
- Deployment FrequencyDeployment FrequencyHow often a team successfully deploys to production. Measures the cadence of shipping.
- Change Failure RateChange Failure RatePercentage of deployments that result in degraded service or require remediation (a hotfix, rollback, or patch). Measures release quality.
- Mean Time to RecoveryMean Time to RecoveryAverage time to restore service after an incident or failed deployment. Measures operational resilience.
All four answer one question: how fast and how reliably does code reach production?
The dominant cost in the change pipeline in the agentic era is no longer commit-to-prod.
It’s the week between “we agreed” and “the first commit lands.”
This guide introduces Intent Lead Time (ILT)Intent Lead Time (ILT)The elapsed time from a product decision to the first commit implementing it. The metric DORA doesn’t measure. Companion to Lead Time for Changes.: the time from a product decision being made to the first commit implementing it. ILT ends exactly where DORA’s clock starts. Together, the two metrics measure the full idea-to-production pipeline.
The time before lead time for changes
For fifteen years, the answer to that question only moved one direction. CI/CD compressed deploys from weeks to minutes. Trunk-based development collapsed merge cycles. Feature flags untangled release from deploy. By 2024, elite teams shipped to production hourly without ceremony.
Coding agents collapsed implementation from days to hours. CI/CD had already collapsed deploys to minutes. The intent-capture phase didn’t collapse at all. Meetings, decision documents, ticket triage, alignment threads: all still on the same clock as ten years ago.
DORA can’t see it. By design.
Yesterday’s assembly line
Software shipped through a human-mediated assembly line. Each handoff was a translation. The PRD imperfectly captured the meeting. The ticket imperfectly captured the PRD. The commit imperfectly captured the ticket. By the time code reached review, the original intent had been refracted through five formats and three job functions, and the engineer who finally implemented it was guessing how the call had actually been made two weeks earlier.
The losses were enormous. The workflow was justified by where the bottleneck lived: in writing the code. Implementation was the long pole. Everything else was overhead in service of getting to a clean commit.
When coding was expensive4, the upstream handoff (the ideation, prioritization, and decision-making) was free. Nobody kept receipts.
Today’s agent-augmented workflow
Then prototyping became a prompt.
Lovable, v0, Cursor, Claude Code, Codex, GitHub Copilot. All of them collapsed implementation from days to minutes. The “AI prototype” lets a team start from working code instead of a blank page.
But the workflow shape is unchanged. Tickets, branches, PRs2, QA gates. The pipeline is still shackled to scaffolding designed for a slower implementation phase that no longer exists.
Worse: the speedup made the upstream loss bigger. Agents execute without context or memory of why. When they ship faster, they ship faster on the wrong intent. The engineering leader’s nightmare has shifted from “we’re shipping too slowly” to “we’re shipping the wrong thing too quickly.”
What DORA measures, and what it doesn’t
DORA is precise about what it measures and where its clock starts.
Lead Time for Changes is defined as the elapsed time from a code change being committed to the repository to the same change running in production. The clock starts at commit. It does not start at decision, at PRD, at ticket, or at “let’s do this.”
This is not an oversight. DORA measures the deployment pipeline because the deployment pipeline is what CI/CD, infrastructure, release engineering, and operational practice can directly improve. Naming the upstream pipeline as DORA’s responsibility would have made the metric unmeasurable for most teams in 2014, when the framework was published.
The asymmetry today is that DORA’s clock starts where the hardest work used to begin. It now starts where that work ends.
Coding is no longer the long pole. The long pole is everything that happens before a single line of code gets written. That is where the next metric lives.
Introducing Intent Lead Time
DORA defines Lead Time for Changes as the time from code commit to code running in production. Intent Lead Time (ILT) is the time from a product decision being made to the first commit implementing it.
ILT = t(first commit) − t(product decision captured)
ILT ends exactly where DORA’s clock starts. The two metrics dovetail at the commit. Together, they measure the full change pipeline, from the moment a decision is made to the moment it ships.
No overlap. No gap.
Four sub-components
ILT is a single number, but it’s composed of four observable handoffs. Profiling each segment shows where time actually lives.
| Sub-metric | Measures | What slows it down |
|---|---|---|
| Capture latency | decision made → decision recorded as an artifact | No live capture. Post-meeting writeups. “We’ll document it later.” |
| Sequencing latency | artifact exists → ticket created | PM triage queue. Sprint planning cadence. |
| Pickup latency | ticket created → assigned | Prioritization backlog. Availability. |
| Activation latency | assigned → first commit | Spec ambiguity5. Context rebuild. “Wait, what were we actually deciding?” |
A healthy ILT compresses all four segments from weeks to hours. A broken ILT is the pipeline everyone has but no one measures.
Three of these are vestigial
Honest version: only one of the four sub-components is load-bearing in an intent-driven workflow. The other three are scaffolding from the assembly-line era.
Tickets are a project-management technology designed for queueing work to humans. The picker needed to be told which one was theirs, when it was their turn, and what state to move it to when they were done. Sequencing, pickup, and activation latencies exist because routing-to-a-person-via-Jira is the only way the old world knew to assign work.
When the picker is an agent, the spec is the assignment. A markdown PRD with acceptance criteria, a design card, an RFC: picked up directly. Nothing to queue, nothing to route, no card to drag across a board.
In a truly intent-driven workflow, sequencing, pickup, and activation will collapse to zero.
That is the provocation, and it is also the prediction. The teams that figure this out first will spend the next decade explaining to everyone else why their tickets-per-engineer counts are dropping while their throughput is climbing.
Proposed bands
A note on these numbers
These bands are first-principles estimates, not benchmarked data. No tool currently measures ILT; these numbers are starting points for the community to refute, refine, or replace. A metric lives or dies by what gets measured against it.
| Band | ILT | Where teams live |
|---|---|---|
| Elite | < 1 hour | Intent captured live. Agents active in the decision session. |
| High | 1 hour → 1 day | Structured specs. Same-day ticket flow. |
| Median | 1 → 3 weeks | Meeting → doc → ticket → assign → commit. The default. |
| Low | > 4 weeks | Decisions surface in Slack threads and never get structured. |
How ILT relates to DORA
DORA’s four keys remain correct for what they measure. Intent Lead TimeIntent Lead TimeThe elapsed time from a product decision to the first commit implementing it. The metric DORA doesn’t measure. Companion to Lead Time for Changes. is a companion metric, not a replacement. It measures what happens before DORA’s clock starts.
Said the other way: DORA compressed deployment from days to minutes. Agents compressed coding from days to hours. Neither moved the clock on intent capture. That is the window ILT opens, and the one where most teams are losing their agentic productivity gains.
Why you probably can’t measure it today
The reason no one tracks Intent Lead TimeIntent Lead TimeThe elapsed time from a product decision to the first commit implementing it. The metric DORA doesn’t measure. Companion to Lead Time for Changes. is not that no one wants to. It’s that the upstream side of the equation has no timestamp.
DORA is measurable because both endpoints are timestamped events in systems engineering teams already use. The commit timestamp lives in git. The deploy timestamp lives in CI/CD. Plumb the two together and the clock writes itself.
The decision timestamp doesn’t exist anywhere structured. It lives in:
- A meeting nobody recorded.
- A Slack thread that scrolled off three days later.
- A PRD written days after the call it captures.
- A whiteboard photo nobody re-types.
- An RFC that landed approved without a clear “this is the moment we agreed” event.
Without a structured decision timestamp, ILT is a number you can argue about but not measure. The teams who do feel like they ship fast usually have a strong written culture that implicitly timestamps decisions: a habit of “decisions go in #decisions on the day they’re made,” or a lightweight RFC process. These are workarounds, not infrastructure.
The metric is unmeasurable until the intent shaping a product decision becomes a first-class object with a clock attached to it.
The intent-driven workflow
Tomorrow’s pipeline doesn’t look like an assembly line. It looks like a continuous human-driven conversation that produces artifacts.
That conversation needs infrastructure: a place for it to happen, a way to timestamp what gets decided, and a handoff-free path from decision to code. This is what Stoa is.
How Stoa makes it concrete
Every abstract piece of the intent-driven workflow maps to a specific product surface in Stoa.
| In the workflow | In Stoa |
|---|---|
| Meetings | Live sessions with video, transcription, and multi-agent participation. |
| Conversations | Real-time transcripts with topics synthesized as you talk. |
| Decisions | Extracted live by AI, timestamped at the moment made, with owner and source attribution. |
| Specs (.md) | PRDs, user stories, and specs drafted live in markdown. No post-meeting writeup. |
| Agentic build | Per-participant Claude Code agents in cloud sandboxes3 with full session context. |
The shape of work changes
The shape of work changes from a relay race to a continuous conversation. Specs become self-documenting because they were captured at the source. The handoff disappears because there is nothing to hand off. The intent behind decisions is already structured, attributable, and machine-readable when the agent picks it up.
This is the workflow that makes Intent Lead Time go from weeks to hours. It is also the workflow that makes ILT measurable. For the first time, the decision event has a timestamp.
DORA gave us the metrics for the DevOps era. Intent Lead Time is the metric for the agentic era.
The clock starts the moment your team agrees.
Author’s note
Intent Lead Time is a time metric. It measures how fast a decision becomes code. It does not measure whether the decision was right.
Software has always had what I’ve called the intent gap: the lossy translation between what we meant and what we built. AI compresses the time, not the gap. A team can have a one-hour ILT and still ship the wrong thing fast.
ILT is necessary but not sufficient. In the fullness of time, it will have a companion metric for decision quality. One that measures whether the intent we captured was the intent we should have captured. I don’t have a name for it yet. But in the agentic era, when coding is no longer the bottleneck, decision quality is the dominant constraint, and someone is going to have to measure it.
Footnotes
- 1.DORA’s framework has since expanded to five metrics: Reliability was added in the 2021 State of DevOps report, and Rework Rate was surfaced in the 2025 State of AI-Assisted Software Development. The “four keys” remain the canonical core; we use that shorthand throughout. ↩
- 2.See “Stop using pull requests” for a sharper version of the argument that PRs are scaffolding from a slower implementation era and now act as drag rather than quality gate. ↩
- 3.Cloud sandboxes are convenient for in-session use, but the build can (and often will) happen outside Stoa. The
stoaCLI gives teams local control: agents run in your own environment, connected to your own infrastructure, with the same captured session context. ↩ - 4.See Dave Griffith, “Software Development in the Time of Strange New Angels”, on how AI coding agents invert the economics of implementation: cheap output, expensive orchestration. ↩
- 5.Spec ambiguity is the inverse of what agile teams call a Definition of Ready: a ticket meets DoR when it’s clear enough to start. Most tickets don’t. ↩
Sources & further reading
- Beyond Code-Centric (SpecStory whitepaper, 2025)
- DORA: Software delivery performance metrics
- 2025 DORA State of AI-Assisted Software Development (Google Cloud)
- The SPACE of Developer Productivity (ACM Queue)
- DevEx: What Actually Drives Productivity (Noda et al., ACM Queue)
- Introducing the DX Core 4 (Abi Noda)
- Project to Product / Flow Framework (Mik Kersten)
- 50 Shades of Lead Time (SourceLevel)