GROWTH|WISE
Opinion

DORA Metrics Tell You What Broke. Coordination Metrics Tell You Why.

DORA is the industry standard for measuring delivery performance. But delivery is an output. The coordination layer upstream of the pipeline determines whether DORA numbers will be healthy or degraded, and DORA doesn’t measure it.

By Growth Wise Research Team 8 min read

DORA metrics earned their authority for a good reason. Google’s DevOps Research and Assessment team built an empirically grounded framework that correlates specific delivery practices with organizational performance. Deployment Frequency, Lead Time for Changes, Change Failure Rate, and Failed Deployment Recovery Time (formerly MTTR) gave engineering leaders a shared vocabulary and a benchmark that actually holds up across industries. The 2021 addition of Reliability as a fifth metric broadened the frame further. Elite, High, Medium, Low: the tiers are clean, the benchmarks are clear, and the framework has genuine predictive power for delivery outcomes.

DORA measures the delivery pipeline: how fast code moves from commit to production and how stable it is once it gets there. The gap is upstream. The human coordination layer that feeds that pipeline (the meetings where scope gets decided, the cross-functional handoffs where ownership gets assigned, the architectural discussions where trade-offs get evaluated) sits outside DORA’s frame entirely. These upstream coordination events determine the quality of what enters the pipeline. DORA picks up the signal after the damage is done.

The gap, metric by metric

Deployment Frequency measures how often code ships. Elite teams deploy multiple times per day. But deployment frequency is partly a function of how cleanly work gets scoped and delegated upstream. When a planning meeting produces vague requirements with no named owner for ambiguous edge cases, the resulting tickets generate blocking questions mid-sprint. Engineers pause, seek clarification, wait for responses that arrive late or contradictory. The deployment that could have shipped Tuesday ships Friday, or gets split into two smaller deployments because the original scope was never properly bounded. DORA registers the slowdown. It can’t tell you it traces back to a meeting where delegation flow was low: actions left the room without enough specificity (owner, next step, deadline) to execute cleanly.

Lead Time for Changes tracks time from commit to production. Elite performers do this in under a day. But lead time includes code review, and code review is where cross-functional coordination failures surface as technical friction. A PR that touches a shared API boundary requires sign-off from another team. If the architectural decision about that boundary was never formally closed (it was discussed in a meeting, opinions were shared, but no one recorded a decision with an explicit owner), the review stalls. The reviewer asks questions the author thought were already settled. A second meeting gets scheduled. Lead time extends by days for a reason that has nothing to do with code quality or CI/CD configuration. The cause is upstream decision reliability: the original decision didn’t close with enough structural completeness to survive contact with implementation.

Change Failure Rate measures deployments that break production. Elite teams keep this under 5%. Some change failures are pure engineering problems: a missed test, a race condition, an untested edge case. But a significant category of production failures traces back to coordination failures that happened weeks earlier. A feature was built against requirements that had silently shifted because two stakeholder groups gave conflicting direction and nobody surfaced the disagreement. An integration broke because Team A changed their API contract without informing Team B, because the meeting where cross-team dependencies were supposed to be reviewed had devolved into a status update with no actual coordination. DORA counts the failure. The root cause is coordination quality: the meeting that was supposed to align these teams didn’t produce real alignment across the structural dimensions that mattered.

Failed Deployment Recovery Time measures how long it takes to restore service. Elite teams recover in under an hour. Recovery speed depends partly on monitoring and incident response tooling. But it also depends on whether the team knows who owns the decision to roll back versus push a hotfix, who has authority to approve an emergency change, and whether the escalation path is clear. When those ownership questions were never explicitly resolved (they were assumed, or distributed vaguely across a group), incident response slows. The team spends the first 20 minutes of an outage figuring out who should be making decisions rather than making them. The upstream signal is coordination debt: unresolved ownership questions that accumulated over prior meetings and now compound under pressure.

The efficient team building the wrong thing

There’s a scenario DORA cannot detect at all. A team deploys frequently, with short lead times, low failure rates, and fast recovery. By every DORA benchmark, they’re Elite. But they’re building features that don’t match what the business actually needs, because the decisions feeding the pipeline were made in meetings where the right stakeholders weren’t present, where dissent was suppressed, or where the stated priorities had drifted from the executive team’s actual intent without anyone making the shift explicit.

This is the most expensive failure mode in product engineering: a high-performing delivery team efficiently shipping work that doesn’t matter. DORA can’t see it because DORA measures the pipeline, not the judgment calls that determine what enters the pipeline. You need a different instrument for that.

Leading indicators, trailing indicators

The relationship between coordination metrics and DORA metrics is temporal. Coordination failures happen first. DORA degradation follows, usually one to three weeks later, when the poorly scoped work, the unclosed decisions, and the vague delegations work their way through the development cycle and hit the deployment pipeline.

This makes coordination metrics leading indicators for DORA. If Coordination Quality drops in a cross-functional planning meeting, you can predict that Lead Time for Changes will extend in the following sprints. If Delegation Flow is consistently low (actions leaving meetings without named owners, deadlines, or clear next steps), Deployment Frequency will slow as engineers spend more time seeking clarification than writing code. If Coordination Debt is rising (the same topics cycling back from prior meetings without resolution), Change Failure Rate will eventually spike because the ambiguity that should have been resolved upstream is now being resolved in production.

The reverse also holds. When coordination metrics are healthy, DORA metrics tend to take care of themselves. Teams that close decisions with explicit owners and documented rationale produce cleaner specs. Cleaner specs produce smaller, more focused PRs. Smaller PRs move faster through review, deploy more frequently, and fail less often. The delivery pipeline accelerates because the coordination layer is feeding it well-defined, properly scoped work.

DORA metrics measure the delivery pipeline. Coordination metrics measure the human decision layer that feeds it. The first set tells you what happened. The second set tells you why. Measuring only DORA is like monitoring server uptime without monitoring the deployment process: you’ll know when things break, but you won’t see the cause until it’s too late to prevent it.

What this means for engineering leaders

DORA isn’t wrong. The framework earned its place as the standard for delivery measurement, and the benchmarks hold. The argument here is additive: DORA is necessary but incomplete. Engineering leaders who instrument only the pipeline see symptoms. Leaders who also instrument the coordination layer see root causes.

For DRIs specifically, this matters acutely. The DRI is increasingly the person held accountable for DORA outcomes on their initiative. When Deployment Frequency drops or Change Failure Rate spikes, the DRI is the one in the room explaining what went wrong. If their only diagnostic tool is DORA itself, they’re stuck pointing at pipeline telemetry that shows the effect but not the cause. Coordination metrics give the DRI the language and the data to trace a delivery failure back to the meeting where the scope was poorly defined, the decision was left open, or the delegation was vague. That shifts the conversation from “the deployment failed” to “the deployment failed because the cross-team dependency review on March 3rd didn’t produce a decision about the API contract, and the ambiguity propagated through two sprints of development.”

One diagnosis leads to a hotfix. The other leads to a structural improvement in how the team coordinates. DORA measures the delivery pipeline. Coordination metrics measure the human layer that determines what the pipeline receives. Both are necessary. Neither is sufficient alone.

Common questions

What are DORA metrics?

DORA metrics are four key performance indicators developed by Google’s DevOps Research and Assessment team: Deployment Frequency (how often code ships to production), Lead Time for Changes (time from commit to production), Change Failure Rate (percentage of deployments causing failures), and Failed Deployment Recovery Time (how long it takes to restore service after a failure). In 2021, DORA added Reliability as a fifth metric. These benchmarks categorize teams into Elite, High, Medium, and Low performance tiers and have become the industry standard for measuring software delivery performance.

What don’t DORA metrics measure?

DORA metrics measure delivery pipeline outputs: speed and stability. They do not measure the upstream human coordination that determines those outputs. A spike in change failure rate might be a CI/CD configuration problem, or it might be that the original decision about what to build never properly closed, leaving engineers working against ambiguous requirements. DORA tells you the deployment failed. It does not tell you that the failure traces back to a meeting three weeks earlier where no one recorded who owned the final architectural call.

How do coordination metrics complement DORA?

Coordination metrics sit upstream of DORA and function as leading indicators. Coordination Quality measures whether cross-functional meetings produce real alignment across structural dimensions. Decision Reliability measures whether decisions close with enough structural completeness to survive pressure. Delegation Flow measures the probability that delegated actions will actually execute. Coordination Debt tracks unresolved items accumulating from prior meetings. When these upstream signals degrade, DORA metrics follow within weeks.

Can a team have elite DORA scores and still have coordination problems?

Yes. A team can deploy frequently with low failure rates and short lead times while consistently shipping the wrong priorities because the decisions feeding the pipeline were never properly validated with stakeholders. DORA measures delivery execution. It does not measure whether what was delivered matched what the organization actually needed. Teams with elite DORA scores and poor coordination quality are efficient at building things that may not matter.

Sources

DORA Team (Google), “DORA’s software delivery performance metrics,” dora.dev/guides/dora-metrics.

Accelerate: The Science of Lean Software and DevOps, Nicole Forsgren, Jez Humble, and Gene Kim. The foundational research behind DORA metrics.

LinearB, “How to improve and measure developer experience.” Developer Experience Index (DXI) and cycle time benchmarks.

Axify, “Measuring Engineering Productivity: From Visibility to Decisions in.” Relationship between visibility and engineering outcomes.

GetDX, “Software development metrics: How to track what really drives performance.”

Stay close to the research

New articles on coordination dynamics, decision reliability, and the science of how teams actually work.

Subscribe to our newsletter