LOBUS.WORKS
← All insights
Flow & Measurement

When Teams Game Metrics: What Anomaly Detection Reveals About Culture

22 Jan 2026·10 min

When leaders discover teams are gaming metrics, the first reaction is usually moral.

People are cheating. The data is dirty. The dashboard cannot be trusted.

That reaction misses the useful part.

Metric gaming is rarely just a data problem. It is a structural signal. It tells you what people believe they must do to survive inside the system you gave them.

The pattern behind the pattern

Teams do not usually wake up wanting to manipulate statuses.

They learn, often very quickly, what the organisation actually rewards. If the official message is "tell us the truth" but the lived message is "do not miss the target," the metrics will start lying on behalf of the team.

That is why anomaly detection is so interesting. The patterns look technical. The story behind them is organisational.

Identity swapping often means work is continuing under a different issue so the old one can close cleanly. Boundary gaming near sprint or month-end usually means reporting windows matter more than flow. Stop-the-clock abuse means excluded statuses are doing political work, not operational work. Selective reporting means somebody learned that appearing healthy is more valuable than being understood.

These behaviours are rational inside a badly designed incentive structure.

Warning

When teams game metrics, they are telling you how the system really works.

What the anomalies reveal

Some patterns point to pressure.

Rapid status transitions, bulk updates, and sudden movements near reporting cutoffs usually show a system that values the appearance of smooth flow more than the reality of it.

Some point to weak psychological safety.

Done-then-reopen loops, follow-up tickets that quietly carry unfinished work, and dummy issues used to dilute ugly distributions all suggest that people do not feel safe leaving complexity visible where leadership can see it.

Some point to power.

Workflow bypasses, field manipulation, and selective reporting usually mean somebody has authority to change the measurement surface without changing the underlying work. That is not a process smell. It is a decision-rights smell.

And some point to governance failure.

Automation-to-done abuse, timestamp mismatches, and assignee or project movement near reporting windows show a system with weak controls around what counts as evidence.

Why the technical detail matters

The strength of anomaly detection is that it takes these stories out of the realm of vague suspicion.

You can see clone chains created around closure. You can spot bursts of transitions that no human workflow would normally produce. You can compare resolved timestamps to actual state changes. You can measure reopen rates, spillover rates, exclusion rates, and automation-heavy completions. You can watch the same game recur at the same boundary every month.

Now the conversation changes.

It is no longer "I have a feeling this team is hiding work." It becomes "this reporting system creates repeated pressure to make work look finished before it is actually finished."

That is a better conversation because it directs attention to system design instead of moral theatre.

The mistake leaders make next

They tighten surveillance.

That feels sensible. Add controls. Audit more fields. Watch people harder. Punish obvious offenders.

Sometimes a control change is necessary. But if the underlying incentive structure stays intact, tighter surveillance usually just produces more creative gaming. The team still has the same problem. It just has to solve it with better camouflage.

The stronger move is to ask what the anomaly is protecting.

What happens to a team that reports the truth? What target, commitment, or political expectation becomes dangerous if the real state of work stays visible? Which leader benefits from the current reporting fiction? Where are people being asked to satisfy two incompatible demands at once?

Read the anomaly as a sentence

Each gaming pattern is saying something.

Identity swapping says, "Work is not really done when the system needs it to be done."

Status theatre says, "We are measured on movement, so we will manufacture movement."

Stop-the-clock behaviour says, "We need somewhere to hide waiting because waiting is politically inconvenient."

Selective reporting says, "The official dashboard is a performance, not a diagnostic."

When you read the anomalies that way, they stop looking like isolated bad acts and start looking like organisational language.

What to do instead

Use anomaly detection as an inquiry tool.

Bring the pattern into the room without accusation. Ask which pressures make it coherent. Compare reported flow with the lived experience of the team. Revisit the targets that taught people to prefer a clean chart over an honest one. Tighten controls only where the measurement surface is too easy to distort, and pair that with changes to the incentives that made distortion useful.

Most of all, stop calling these problems data quality issues as if the data somehow got corrupted by itself.

The data is telling you about power, safety, and adaptation.

That is exactly why it matters.