KPI Tree
KPI Tree

From set-and-forget to structured monitoring

How to track OKR progress with metric trees

Most OKR programmes fail not at the goal-setting stage but at the tracking stage. Teams write ambitious objectives in January and discover in March that nobody checked whether they were on course. This guide covers the check-in cadence, scoring methods, and diagnostic techniques that turn OKRs from quarterly wish lists into a genuine performance system, with metric trees as the structural backbone.

9 min read

Generate AI summary

Why OKR tracking fails

The core problem

The number one reason OKR programmes fail is not poor goal-setting. It is the set-and-forget pattern: objectives are written at the start of the quarter, pinned to a wiki page, and never revisited until the quarter-end retrospective. By then, the data is a post-mortem, not a navigation tool.

Research from OKR benchmark studies consistently shows that teams who update their key results weekly are over 40% more likely to achieve their objectives than teams who check in monthly or less. Yet the majority of organisations treat OKR tracking as something that happens at the end of the quarter rather than throughout it.

The reasons are structural, not motivational. People do not ignore their OKRs because they are lazy. They ignore them because the tracking process adds overhead without delivering insight. A typical OKR check-in involves opening a spreadsheet, updating a percentage, and writing a status comment that nobody reads. There is no diagnosis, no context, and no clear path from "this number is behind" to "here is what we should do about it." The check-in ritual becomes performative rather than useful, and performative rituals get abandoned.

The second failure mode is treating all key results as equally informative. A key result might show 60% completion at the midpoint of the quarter. Is that good or bad? It depends entirely on what is driving that number and whether the trajectory is improving or degrading. Without a structural model that shows how the key result connects to its underlying drivers, teams are left staring at a number with no way to interpret it. They know they are behind but they do not know why, and without a why, they cannot act.

Set and forget

OKRs are written once and revisited only at the quarter-end review. Without a regular cadence, they drift from daily work and become administrative artefacts rather than operational tools. The longer the gap between check-ins, the larger the gap between intent and reality.

Tracking without diagnosis

Teams update the percentage but never investigate why the number moved or stalled. A green status hides the question of whether the underlying drivers are healthy. A red status triggers alarm but not understanding. The check-in answers "where are we?" but not "why are we here?"

No connection to daily work

Key results live in one tool, project tasks in another, and the link between them is maintained only in people's heads. When the connection between what teams do each week and what the OKR measures is implicit rather than explicit, the OKR stops influencing decisions.

Lagging indicators only

Most key results measure outcomes that are visible only after the quarter is well underway. Revenue, retention, NPS, and conversion rates are lagging signals. By the time they indicate a problem, the window for course correction has narrowed or closed entirely.

The check-in cadence that actually works

Effective OKR tracking operates on a layered cadence. The mistake most organisations make is choosing a single frequency and applying it uniformly. In practice, different levels of the organisation need different rhythms, and the check-in itself must serve a different purpose at each level.

The weekly check-in is the heartbeat of the system. It is short, typically fifteen to twenty minutes per team, and focused on signal detection rather than deep analysis. Each key result owner answers three questions: what is the current value, what is your confidence that this key result will be achieved by quarter-end, and what, if anything, is blocking progress? The point is not to solve problems in the meeting. The point is to surface them early enough that they can be solved during the week.

The fortnightly or monthly review goes deeper. This is where the team examines trajectories rather than snapshots. Is the key result accelerating, decelerating, or flat? Are the underlying drivers moving in the right direction? Are there leading indicators that suggest the current trajectory will change? This review typically lasts thirty to sixty minutes and involves looking at the metric tree to understand what is happening beneath the key result.

The mid-quarter review, held around week six of a twelve-week cycle, is the decision point. By this stage you have enough data to assess whether the OKRs are still achievable with the current approach or whether a course correction is needed. This is where you prune initiatives that are not working, pivot to new approaches, or in rare cases adjust the key result target itself if the assumptions behind it have proven wrong.

CadenceDurationPurposeKey question
Weekly15-20 minSignal detection and confidence scoringAre we on track, and are there any blockers?
Fortnightly or monthly30-60 minTrajectory analysis and driver diagnosisWhy is the number moving (or not), and what do the leading indicators say?
Mid-quarter (week 6)60-90 minCourse correction decisionsDo we need to change our approach, reallocate resources, or adjust the target?
Quarter-end60-90 minScoring, learning, and input to next cycleWhat did we achieve, what did we learn, and what carries forward?

The layered cadence prevents two common pathologies. It prevents the set-and-forget pattern by creating regular touchpoints that keep OKRs in working memory. And it prevents check-in fatigue by keeping the weekly ritual lightweight and reserving deeper analysis for less frequent sessions. Teams that try to do deep diagnostic reviews every week burn out on the process. Teams that only review monthly miss problems until they are too large to fix.

Scoring methods compared

How you score OKR progress shapes how people behave. The scoring method is not a neutral measurement choice; it sends a signal about what the organisation values. A binary system rewards completion. A percentage system rewards incrementalism. A confidence-based system rewards honest forecasting. Choosing the right method, or the right combination, matters more than most OKR guides acknowledge.

There is no universally correct approach. The best scoring method depends on the nature of the key result, the maturity of the OKR programme, and the culture of the organisation. What follows is a comparison of the four most common approaches, with guidance on when each is most appropriate.

MethodHow it worksBest forWatch out for
Binary (yes/no)The key result is either achieved or not. No partial credit.Activity-based key results (launch a feature, hire a role, complete a migration) and committed OKRs where partial delivery is not acceptable.Penalises stretch. A team that achieves 90% of an ambitious target scores the same as one that achieved 0%. Can discourage risk-taking if overused.
Percentage completionProgress is measured as a percentage of the target. If the KR is "Increase revenue to £12M" and you reach £11.4M, you score 95%.Quantitative key results with a clear numeric target. The most intuitive method for data-literate teams.Treats all percentages as linear. Getting from 0% to 50% may require fundamentally different effort than 50% to 100%. Does not distinguish between "on pace" and "front-loaded then stalled."
Traffic light (RAG)Key results are rated green (on track), amber (at risk), or red (off track) based on current trajectory and confidence.Executive-level reporting, cross-team alignment meetings, and organisations early in their OKR journey who need simplicity.Subjective without clear criteria. One team's amber is another's red. Must be anchored to specific thresholds (e.g. green = >70% confidence of hitting target) to be meaningful.
Confidence scoringEach week, the KR owner rates their confidence that the key result will be achieved by quarter-end, typically on a scale from 1 to 10 or as a percentage likelihood.Aspirational OKRs where progress is non-linear. Teams that want early warning signals before the numbers themselves deteriorate.Requires psychological safety. In blame cultures, people will inflate confidence scores to avoid difficult conversations. Only works when honest assessment is rewarded.

Many mature OKR programmes use a hybrid approach. Committed OKRs, the ones the team has agreed must be delivered, use binary or percentage scoring because partial completion is not acceptable. Aspirational OKRs, which are intentionally set beyond comfortable reach, use confidence scoring because the goal is to learn and stretch rather than to guarantee delivery. Google famously considers a score of 0.6 to 0.7 on their decimal scale to be a strong outcome for aspirational OKRs. Consistently scoring 1.0 indicates the goals were not ambitious enough.

The traffic light system is most useful as a communication layer on top of another scoring method rather than as a primary scoring mechanism. It works well in leadership reviews where the audience needs to absorb the status of many OKRs quickly. But it should be derived from underlying data, not assigned based on gut feeling. A key result is green when the confidence score is above 7 and the trajectory supports it. It is amber when confidence drops below 5 or when leading indicators suggest a problem. It is red when the team believes the target is unlikely to be met without a significant change in approach.

“The purpose of scoring is not to judge performance. It is to generate an honest signal that enables timely decisions. A red score delivered in week four is worth far more than a red score delivered in week thirteen.

How metric trees make OKR check-ins diagnostic

The most common frustration with OKR check-ins is that they surface problems without explaining them. A key result is behind target, and the team knows they are behind, but the check-in process offers no structured way to understand why. The conversation devolves into speculation: maybe it is the new competitor, maybe it is the feature delay, maybe it is seasonal. Everyone has a hypothesis and nobody has evidence.

A metric tree changes the nature of the check-in from status reporting to diagnostic investigation. When a key result is mapped to a node in the metric tree, the team can immediately trace downward through the tree to identify which sub-drivers are underperforming. This is not a vague analytical exercise. It is a specific, repeatable navigation: the key result node shows a shortfall, so you look at its child nodes to find the source.

Consider a SaaS company tracking the key result "Increase monthly recurring revenue to £850K." The metric tree decomposes MRR into its structural drivers. During the week-six check-in, MRR is at £790K against a target trajectory of £815K. Rather than debating why, the team navigates the tree.

The tree reveals that new MRR is the underperforming branch. Drilling deeper, trial starts are on target but the trial-to-paid conversion rate has dropped from 12% to 9%. One level further, onboarding completion percentage is stable but time to first value has increased, suggesting that recent product changes have made the initial experience slower. In under five minutes, the team has moved from "MRR is behind" to "the onboarding flow after the recent release is adding friction that delays value delivery, which is reducing trial conversion."

This diagnostic capability transforms the quality of the check-in conversation. Instead of debating hypotheses, the team is discussing a specific, evidence-based finding. Instead of assigning blame, they are identifying a lever. And instead of a vague action item like "investigate MRR shortfall," they have a precise one: "review the onboarding flow changes from the last release and measure their impact on time to first value."

In KPI Tree, this navigation happens visually. Each node in the tree shows its current value, target, and trend. Nodes that are off track are highlighted, creating an instant visual map of where the problem is. Teams can click through the tree during the check-in itself, making the diagnosis a collaborative exercise rather than a homework assignment that someone does between meetings.

Connecting key results to leading indicators

Most key results are lagging indicators. Revenue, retention, NPS, and market share all reflect outcomes that have already happened. By the time a lagging indicator signals a problem, the damage is done and the quarter is half over. Effective OKR tracking requires pairing each lagging key result with leading indicators that predict its trajectory before the outcome materialises.

A metric tree naturally exposes this relationship. The nodes closer to the root of the tree tend to be lagging indicators: revenue, customer count, profit margin. The nodes further down the branches tend to be leading indicators: pipeline velocity, feature adoption rates, support ticket volume, onboarding completion rates. These lower-level metrics move before the higher-level ones, which is precisely what makes them useful for early warning.

  1. 1

    Identify the key result node in the tree

    Locate where your key result sits in the metric tree. If the key result is "Reduce churn to 3%," find the churn rate node and examine its position relative to the tree's root and leaves.

  2. 2

    Trace downward to find input metrics

    Navigate down the tree from the key result node to identify the operational inputs that drive it. Churn rate might decompose into voluntary churn (driven by product satisfaction and competitor activity) and involuntary churn (driven by payment failures and billing issues). Each of these decomposes further into metrics your teams directly influence.

  3. 3

    Select two to three leading indicators per key result

    Choose the input metrics that move earliest and have the strongest causal link to the key result. For churn, these might be "support ticket volume per account" (a leading signal of dissatisfaction), "feature adoption depth" (accounts using more features churn less), and "payment failure rate" (a leading signal of involuntary churn). These become your early warning system.

  4. 4

    Monitor leading indicators in weekly check-ins

    During the weekly check-in, review the leading indicators alongside the key result itself. If support ticket volume is rising but churn has not yet increased, you have a window to act. If feature adoption is declining in a specific cohort, you can intervene before those accounts churn. The leading indicators buy you time that lagging indicators cannot.

  5. 5

    Set threshold alerts on leading indicators

    Define specific thresholds that trigger investigation. If support tickets per account exceed 2.5 per month, or if feature adoption drops below 60% for any cohort, the team investigates regardless of what the lagging key result currently shows. These thresholds function as trip wires that catch problems while they are still small enough to fix.

Practical tip

In KPI Tree, you can set alerts on any node in your metric tree. Configure leading indicator nodes to notify the key result owner when thresholds are breached. This turns passive tracking into active monitoring, so that problems surface in your inbox rather than waiting for the next scheduled check-in.

What to do when OKRs go off track mid-quarter

An OKR going off track is not a failure. It is information. The question is what you do with that information, and the answer depends on the diagnosis. Too many teams respond to a red key result with either panic (throw resources at the problem) or resignation (accept the miss and move on). Both responses waste the learning opportunity that the variance represents.

The structured response follows a decision tree. First, diagnose. Then decide whether to adjust the approach, adjust the target, or accept the variance and capture the learning.

  1. 1

    Diagnose using the metric tree

    Before deciding what to do, understand why the key result is off track. Navigate the metric tree downward from the key result node to identify the specific driver that is underperforming. Is it a single bottleneck or multiple factors? Is the cause within the team's control or external? Is the underperformance accelerating or stabilising? The diagnosis determines the response.

  2. 2

    Prune initiatives that are not working

    If the metric tree shows that the initiative intended to move the key result is not affecting its target node, stop it. Continuing to invest effort in an approach that the data says is not working is the most common waste in quarterly cycles. Pruning frees up capacity for a different approach.

  3. 3

    Pivot to a different lever

    The metric tree may reveal that the originally targeted driver is harder to move than expected, but an adjacent driver offers a faster path to the same outcome. If improving onboarding completion is proving slow, but reducing payment failures is quick and addresses a meaningful portion of the gap, pivot to that lever. The tree shows you the alternative paths.

  4. 4

    Adjust the target if assumptions were wrong

    Sometimes the original target was based on assumptions that have since proven incorrect. If a key market shifted, a dependency failed to materialise, or the baseline data was inaccurate, it is better to adjust the target to something honest than to maintain a fiction. Adjusted targets should be clearly documented with the reason for the change, so the learning is captured.

  5. 5

    Accept the variance and capture the insight

    Not every off-track OKR needs to be rescued. If the diagnosis shows that the miss is due to factors genuinely outside the team's control, or that the cost of closing the gap exceeds the benefit, accept the variance. The critical step is to document what was learned. Why did the plan not work? What assumption was wrong? What would you do differently? This learning feeds directly into the next OKR cycle and into the metric tree itself, making both more accurate.

The mid-quarter review (around week six) is the natural decision point for these course corrections. By week six, you have enough data to distinguish between a temporary dip and a structural problem, but you still have enough time in the quarter to act meaningfully. Organisations that wait until week ten or eleven to address off-track OKRs are left with too little runway for any intervention to take effect.

Psychological safety is the prerequisite for all of this. If raising a red flag results in blame or punishment, people will not raise red flags. They will mask problems, inflate confidence scores, and present an artificially green picture until the quarter end forces a reckoning. The entire tracking system depends on honest signals, and honest signals depend on a culture where surfacing problems early is rewarded rather than penalised.

“A red key result surfaced in week four, diagnosed through the metric tree, and addressed with a pivot by week five is a success story. A red key result hidden until the quarter-end retro is an organisational failure, not an individual one.

Putting it all together

OKR tracking is not a single practice. It is a system composed of cadence, scoring, structural diagnosis, leading indicators, and course correction. Each element reinforces the others. The weekly cadence surfaces signals early. The scoring method generates honest assessments. The metric tree provides the diagnostic depth to understand what the signals mean. Leading indicators extend the warning window. And the course correction framework turns understanding into action.

Organisations that implement all five elements report dramatically better OKR outcomes, not because the goals themselves change, but because the feedback loop between intent and execution becomes tight enough to actually steer by. The OKRs stop being a quarterly planning exercise and become the operating rhythm of the business.

Weekly signal detection

Every key result owner updates their confidence score and current value weekly. The check-in is fifteen minutes, focused on surfacing blockers and changes in trajectory. This rhythm keeps OKRs in active working memory.

Tree-based diagnosis

When a signal indicates a problem, the team navigates the metric tree to identify the root cause. The tree turns a vague "we are behind" into a specific "this driver in this branch has dropped because of this factor." Diagnosis takes minutes, not weeks.

Leading indicator monitoring

Each key result is paired with two to three leading indicators drawn from lower nodes in the metric tree. Threshold alerts on these indicators provide early warning before the lagging key result itself deteriorates.

Structured course correction

At the mid-quarter review, off-track OKRs are addressed through a clear decision framework: diagnose, prune, pivot, adjust, or accept. Every decision is documented, and the learning feeds back into the metric tree and the next OKR cycle.

The metric tree is the connective tissue that makes all of this work. Without it, weekly check-ins are just status updates. Scoring is just number tracking. Course correction is just guesswork under pressure. The tree provides the structural context that transforms each practice from a ritual into a tool. It shows why numbers are moving, which levers to pull, and what the downstream consequences of any change will be.

If your organisation already uses OKRs, you do not need to overhaul the framework. You need to add the structural layer that makes tracking genuinely diagnostic. Build a metric tree, map your key results onto it, and watch how the quality of your check-in conversations transforms from "are we on track?" to "what is actually happening in the system, and what should we do about it?"

Stop guessing why your OKRs are off track

KPI Tree connects every key result to the metric tree that explains it. See which drivers are underperforming, get early warnings from leading indicators, and make your weekly check-ins genuinely diagnostic.

Experience That Matters

Built by a team that's been in your shoes

Our team brings deep experience from leading Data, Growth and People teams at some of the fastest growing scaleups in Europe through to IPO and beyond. We've faced the same challenges you're facing now.

Checkout.com
Planet
UK Government
Travelex
BT
Sainsbury's
Goldman Sachs
Dojo
Redpin
Farfetch
Just Eat for Business