Decision Escalations: The Ladder That Keeps Work Moving
When escalation is driven by feelings, decisions become political. Design the trigger, and the system can move.
The line is running.
Nothing’s flashing red. Nobody’s panicking. If you walked by, you’d assume the system is healthy because the system is moving.
But you can feel it. Drift. The kind that doesn’t announce itself as a failure. It shows up as small adjustments that start to feel normal. Tooling gets tweaked more often than it used to. Operators compensate in slightly different ways. The first hour of the shift never really settles. The process still produces output, but it takes more “touch” to keep it there.
This is where escalation gets misunderstood.
Most organizations treat escalation like a breakdown. Like someone couldn’t handle their scope, so now leadership has to get involved. That framing trains the worst behavior. People hide problems until the cost is high enough that nobody can ignore it, then they call it urgency and pretend it arrived out of nowhere.
Escalation isn’t a failure. It’s a feature. It’s how a system protects itself from drift.
Here’s our thesis: if escalation is driven by feelings, decisions will become political. If escalation is undefined, autonomy either freezes or drifts. The fix is simple and hard at the same time. Tie escalation to measurable impact. Route it through a clear ladder of authority. Put a timebox on the decision. Record what you chose and what you learned.
Not because you love process. Because you want movement you can trust.
On the floor, you’re seeing the early warning signs in tooling and operator variation. And it matters that you’re not claiming to be the quality subject matter expert. You’re seeing system behavior. The line lead is making judgment calls with incomplete feedback. Operators are adapting in real time. Tooling wears in subtle ways, and variation stays inside the acceptable band until the day it doesn’t. When the signals are weak or scattered, escalation becomes a confidence game.
Who calls it?
Who pauses and checks?
Who says, “This drift is now a broader trade-off, not a local adjustment.”
When confidence is the trigger, escalation turns into theater. Not because people are careless. Because we’re asking humans to make statistical calls without the instrumentation, visibility, and shared thresholds to support those calls.
The thing manufacturing gets right, often without ever naming it, is that escalation has to exist. You don’t ship governance decks on a truck. You ship product. So even if it’s informal, the ladder exists. Process line leads and manufacturing leadership already play defined roles when things start to slide.
The opportunity is to turn that instinct into something repeatable. Something you can teach. Something that doesn’t depend on who’s on shift.
Now pivot with me, because the data side is usually the mirror image.
In master data, escalation often doesn’t exist yet. Not as a system. Not as a teachable path. Not as something you can audit and improve.
Here is the truth, and it’s the most honest line in this whole chapter: customer setup exceptions don’t have a home yet. For most, if you had to guess where they live today, it’s probably email and word of mouth.
That’s not a moral failing. It’s the default state of early data management.
But it is a shadow operating model. Quiet decisions, unlogged risk, and workarounds that compound until they become “how we do things.” Later, when you try to build customer master data, you discover you’re not just cleaning data.
You’re unlearning an underground system of exceptions that grew in the dark.
Think about your own data organizations; "Customer setup" could be the first use case you’re going after, and it’s the right one because it sits at the seam where friction multiplies. Retailers. Distributors. Ship-to complexity. Hierarchies. Identifiers that have to be correct across systems, not just in one place. When customer master data is wrong or incomplete, the business doesn’t stop. It routes around it. Someone emails someone. Someone creates a one-off. Someone uses an alias. Someone will “fix it later.”
That’s fine once.
At scale, those workarounds become the operating model. Orders get held. Shipments get delayed. EDI breaks. Pricing conditions don’t apply. Two “different” customers look identical everywhere except the place that matters. Now you’re not troubleshooting a data issue. You’re troubleshooting trust.
These two worlds are closer than they look. Tooling drift and customer setup exceptions share the same underlying pattern: the system can’t see the deviation early enough, so people compensate late. They do the best they can with what they have, but the organization pays for it in rework, delay, and a slow erosion of confidence.
Escalation design matters. It is the bridge between autonomy and reliability.
Here’s the operating pattern I keep coming back to because it stays usable under pressure.
Trigger. Owner. Timebox. Target. Record.
It's not a committee. Not a vibe.
A designed path.
- A trigger tied to impact beats a room full of opinions.
- A ladder isn’t about control. It’s about routing decisions to the right level.
- A timebox prevents slow death by discussion.
- A record turns heroics into learning.
Let’s make it concrete.
On the manufacturing side, you don’t need perfect instrumentation to move away from feelings. You need consistent signals:
Scrap rate. Yield. Defect codes. Rework events. How often tooling is being adjusted mid-shift. What happens after changeover.
Even if the data isn’t clean yet, it can still be directional enough to support a trigger.
A practical trigger is "baseline plus drift". Something like: scrap exceeds baseline by a meaningful margin for a sustained window. Or the same defect repeats enough times that you’re no longer looking at random noise. Or tooling interventions exceed a threshold that tells you the process is being held together by constant adjustment.
Then you assign ownership.
On the floor, Rung 1 belongs with the process line lead. They should be able to act quickly inside guardrails. Pause briefly. Run a defined check. Contain the last set of units. Reset what can be reset. That’s autonomy doing its job.
But you don’t let it linger.
You timebox it. If the line lead can’t return the process to green quickly, you escalate to manufacturing leadership. Not because the line lead failed, but because the decision has changed. It’s no longer a local optimization.
It’s now a trade-off decision with broader impact.
And finally, you record it. Even if the record starts lightweight. Trigger met. Action taken. What changed. What you’re watching next. If it repeats, it becomes corrective action, not folklore.
Now apply the same pattern to our customer setup, where the exception path is likely email and/or word of mouth.
Start with triggers the business actually feels. “Data quality is low” won’t move anyone. “Orders are on hold because customer setup is incomplete” will. So will repeat EDI failures tied to customer identifiers, recurring ship-to creation delays, duplicate customer creation events, or invoicing exceptions caused by broken hierarchies.
Then you assign an owner, even if it’s a beta construct at first. A named data steward, or an on-call steward role, even if the rotation is one person today. The goal is not perfection. The goal is to create a predictable first rung.
Rung 1 needs guardrails. A controlled exception is not “do whatever it takes.” It’s a temporary bypass with an expiration date, a reason code, and a linked remediation task. Otherwise you’ve just formalized the shadow system.
Then you timebox it. If it can’t be resolved quickly, it moves to Rung 2. Cross-functional. The process owner plus the ERP or integration lead plus whoever is carrying governance. This is where you decide fix now, controlled workaround, or business-approved override.
And if the choice is truly “ship now versus data purity now,” it belongs at Rung 3 with an explicit business owner accepting the risk.
That’s not bureaucracy. That’s honesty.
The final piece is the record. If it isn’t logged, it didn’t happen. Not because you love paperwork, but because you want learning. Email is not a system. It’s just where systems go to die quietly.
If you want a tangible artifact to anchor this chapter, it’s a Decision Card. One page. Every time.
Decision name. Why it matters. Trigger. Rung 1 owner and guardrails. Rung 2 escalation target and scope. Rung 3 business trade-off owner. Timeboxes. Decision record minimum.
Two Decision Cards are enough to make this real.
One for tooling and operator drift: process line lead to manufacturing leadership, triggered by sustained drift, timeboxed, recorded.
One for customer setup exceptions: steward to cross-functional to business owner, triggered by business friction, timeboxed, recorded, with temporary exceptions that must close with a permanent fix or an explicit rule change.
This is where your “beta” framing becomes a strength. You’re not pretending this is fully built. You’re showing how decision systems get built in the real world. You design a path, you run it in production, you tighten the guardrails, and you let the work teach you what’s missing.
That’s how systems become muscle.
If you want to start tomorrow without buying anything or launching a massive governance program, do this.
Pick three recurring decisions that currently waste time or create rework. One in your operations. One in a specific domain. One that crosses teams. Write a Decision Card for each. Define the trigger. Name the owner. Put a timebox on it. Define the escalation target. Decide what gets recorded.
Then run it for two weeks and review the path, not just the outcome.
- Where did it stall?
- Where did it escalate too soon?
- Where did it escalate too late?
- Where did the record reveal the same exception pattern repeating?
That loop is the work. Not theory. Not bureaucracy. Operational learning.
Let me leave you with one question:
What’s one decision in your organization that only escalates when it’s already expensive, and what measurable trigger could move that escalation earlier, with a named owner and a clear ladder?
Sources
Deming, W. E. (1986). Out of the crisis. MIT Press.
Reason, J. (1997). Managing the risks of organizational accidents. Ashgate.
Weick, K. E., & Sutcliffe, K. M. (2015). Managing the unexpected: Sustained performance in a complex world (3rd ed.). Wiley.