Every organization keeps score. Some track revenue and burn, others obsess over NPS or uptime, and a few measure everything that moves. Yet when you walk factory floors, sit in standups, or read monthly dashboards, you often find numbers that are easy to collect but hard to use. Metrics turn into wallpaper, leaders chase noise, and teams game what they can’t influence. Deming warned about this long before digital dashboards. His 14 principles, crafted in the crucible of postwar manufacturing and refined through decades of quality work, remain a remarkably practical guide for building metrics that focus effort, protect learning, and deliver real outcomes.
What follows is the approach I’ve used across software, healthcare, and advanced manufacturing to design metrics that actually move the system. Deming gives us the philosophy and the guardrails. The rest is careful carpentry: fit for purpose, built with context, and tuned over time.
The hazard of measuring what is easy
I once worked with a mid-sized SaaS company that set a top-level goal to “increase customer love” and then made NPS the north star. Within weeks, support agents began prompting for surveys after quick fixes, sales delayed sending onboarding emails until after go-live euphoria, and product shipped minor interface tweaks to harvest fast wins. NPS rose 7 points in a quarter. Renewals didn’t budge. Buried deeper in the logs, we later saw a rise in unacknowledged critical bug reports and a drop in adoption of advanced features. The metric was real, and it moved, but the system that mattered did not.
This is the core mistake. We treat the measure as the aim. Deming reversed that. Clarify the aim of the system, then choose measures that teach you whether you are moving closer to it. Not the other way around.
The hinge: constancy of purpose and the aim of the system
Deming’s first principle, create constancy of purpose, is where metric design starts. Define the enduring aim with enough precision that it can guide trade-offs when the numbers flicker. The best organizations I’ve worked with write the aim in language that describes who is served, how they benefit, and what risks are unacceptable. For example, a medical device team framed its aim as “deliver diagnostic results within 30 minutes at 99.5 percent reliability for community clinics, while continuously lowering cost per test.” That statement drove a small, stable metric set: turnaround time distribution, drop rates by clinic type, and blended cost per test. By naming the user and constraints, they avoided chasing vanity numbers like total tests shipped.
Without constancy of purpose, metrics become a bargaining chip. Sales wants top-line growth, operations wants utilization, engineering wants cycle time, finance wants unit economics. Everyone can show a graph that rises. Nobody can prove the whole is getting better.
Work with the system you have, not the one you wish for
Deming’s deep lesson on systems thinking shows up every time a manager asks, “Who caused that defect?” Most variation comes from the system, not from individual workers or single events. Metrics must reflect that reality, or they will drive blame instead of improvement.
If you operate a call center, for example, the natural variation in daily call volume can swamp agent-level differences. A control chart on average handle time by day teaches you when a process change is needed. Ranking agents by last week’s handle time and rewarding the top quartile teaches luck. One organization I coached abandoned weekly league tables and instead built a single system-level run chart with clear process limits. When a special cause appeared, they asked two questions: what changed in the environment, and what did we do that might have created a signal? Their interventions moved from pep talks to process design.
Measure causes you can influence and outcomes you actually seek
Deming cautioned against management by visible figures alone. That often shows up as lagging metrics dominating the story. Revenue, churn, defects found in the field, mortality rates after surgery - these are vital, but they register too late for daily steering. On the other hand, a dashboard filled only with inputs and activities invites busywork. The discipline is to pair outcomes with drivers you can influence, then track the relationships.
In a robotics assembly plant, final test yield was the outcome. The drivers were torque application accuracy, parts kitting error rates, and solder joint temperatures within spec. The team plotted yield alongside these drivers over many weeks. Two patterns mattered. First, yield shifts only when at least two drivers degrade, so single-driver sprints rarely paid off. Second, solder temperatures below the lower control limit preceded yield drops by one to two days. They added a morning check of reflow profiles and prevented three weeks of yield pain with a single maintenance step. The metric set worked because the team understood the mechanism.
Constancy hates knee-jerk reactions: understanding variation
A decision rule turns a number into action. Without a good rule, every dip in a weekly chart sparks a memo and every bump inspires a victory lap. Deming’s teaching on variation, especially through control charts, cuts through the noise.
You don’t need to publish a statistics primer to use this well. In software teams, I’ve used a lightweight approach:
- For core flow metrics like cycle time or change failure rate, chart at least 20 to 30 consecutive periods before deciding if a shift is real. Use simple rules of thumb: eight points in a row above the median suggests a process change, a single outlier more than three standard deviations away needs investigation, and scattered ups and downs around the mean call for patience.
This is one of only two lists in this article, and it reflects practical observation rather than formula worship. The spirit matches Deming’s guidance: build prediction into your measurement. When a number changes, ask whether the underlying process changed or if random fluctuation did its usual dance.
A healthcare operations leader I worked with kept a pocket card with two questions. First, is this common cause variation? If yes, don’t react with an individual fix. Improve the process. Second, is this special cause variation? If yes, look for a specific reason and correct it. That simple split saved her team dozens of wasteful root-cause hunts each quarter.
Break down barriers by making measures shared, not competing
Deming’s principle about breaking down barriers between departments often clashes with metric design. Marketing wants MQLs, sales wants SQLs, product wants activation, support wants CSAT. Separately, each metric is fine. Together, they create friction. The cure is to define a small set of system measures that cross boundaries, and then give teams joint accountability.
In one B2B software company, activation within 14 days sat at 42 percent. Marketing, proud of top-of-funnel growth, was sending prospects who were poor fits. Sales closed deals with heavy discounting but thin customer education. Customer success inherited the mess. We reframed the metric to “customers reaching first value in 14 days,” defined concretely as completing three workflows that predicted long-term retention. Marketing adjusted targeting, sales changed demos to emphasize those workflows, and customer success reworked onboarding. Activation reached 63 percent in two quarters with flat acquisition spend. The single, shared metric created cross-functional focus without the usual blame.
Eliminate slogans, add operational definitions
Deming’s advice to eliminate slogans lands squarely on metrics. “Zero defects” and “delight every customer” sound noble. They also hide choices and invite manipulation. Replace them with operational definitions that can be measured the same way by any reasonable person.
A logistics operation used “on-time delivery” as a hero metric. Depends on who asked. Operations counted a delivery as on time if it reached the local depot by the promised date. Customer service counted the customer’s receipt date. Finance used the invoicing date. No surprise that the numbers didn’t match. After a messy week, the team defined on time as “arrived at the customer’s dock by 5 p.m. on the committed date in the order system.” Overnight, arguments about the number gave way to arguments about the process, which is where energy belongs.
When you create an operational definition, include three things. What is counted and what is excluded, the method of measurement including source systems, and the time boundary. Do that, and the metric becomes a tool rather than a political football.
Drive out fear with measures that do not punish learning
Deming warned, drive out fear. Metrics can either amplify fear or reduce it. The difference lies in how finely you slice and how you use the results. If you publish weekly leaderboards of individual performance, expect sandbagging and stress. If you measure outcomes at the system level and use team-level data to improve, expect candor.
A clinical lab I worked with tracked individual technician error rates and printed red numbers next to names. Rework fell for a month, then rebounded, and morale cratered. When they switched to system-level error rates and invited technicians to annotate the chart with context, patterns emerged. New hire cohorts made the same mistakes in week three. Night shift errors rose after software deployments. Fixes targeted training and deployment timing. Error rates fell and stayed low. The technicians knew they were being measured, but they were not being shamed. That change freed them to share what they saw.
Cease dependence on inspection: build capability, then measure to prove it
Deming pushed hard against inspection as the primary quality strategy. In metrics, inspection shows up as counting defects at the end and trying to read tea leaves. The alternative is to measure the capability of upstream processes and improve them.
In a fintech back office, reconciliation errors were caught weekly. Leadership set targets for “errors found,” congratulating the team when the number was high because “we’re catching more.” That mindset rewarded the symptom. We replaced it with process capability measures: automated match coverage rate, exception queue age, and mean time to reconcile by exception type. The weekly defect count stayed as a lagging check, but team energy went to expanding automation coverage from 68 to 85 percent and reducing long-aged exceptions by 60 percent. Downstream errors naturally fell, and the team stopped glorifying inspection.
Without leadership, metrics degrade into targets and games
Deming put responsibility on management to institute leadership. In metric terms, leadership means curiosity, not command. Ask for the story behind the number, not just the number. Reward teams that surface bad news early. Accept the trade-offs that constancy of purpose demands.
I watched a VP of engineering do this well. Deployment frequency dipped after they introduced an approval gate for high-risk services. Rather than order the team to push more often, he asked for before and after control charts, incident trends, and developer sentiment. The data showed that deployment batch size increased slightly, but change failure rate dropped from 6 to 2 percent, and time to restore improved by 30 percent. He left the gate in place for those services and encouraged the team to find ways to reduce batch size through better test isolation. The metric became a conversation, not a cudgel.
The trap of targets without methods
Deming cautioned that setting numerical goals without methods invites distortion and dysfunction. Still, executives need goals, and teams need to know what good looks like. The fix is simple to say and hard to do: set limits and aspirations, then pair them with a plan to change the system.
If you want customer wait times under two minutes 90 percent of the time, identify how you will achieve that structurally: staffing models, cross-training, improved routing, better self-service. Publish the method with the target. Review both together. When the number misses, ask whether the method was executed, whether the assumptions held, and what systemic constraints stood in the way. When the number hits, verify that you did not degrade something else you care about, like resolution quality. This pairing keeps you honest.
Small experiments, rapid learning
Deming favored improvement by method, not heroics. The best metrics encourage small tests of change, using data to learn quickly. A software team I advised stopped monthly dashboard brawls and started weekly experiments. They would pick a hypothesis, for example, “adding a pre-commit checklist for risky modules will reduce change failure rate without slowing delivery.” They instrumented the modules, ran the checklist for two weeks, and watched a control chart. If the signal showed improvement without increasing lead time materially, they kept it. If not, they rolled back. Over six months, change failure rate halved and lead time stayed in a tight band. The key was not the checklist. It was the habit of pairing measures and experiments.
Respect for people, respect for context
Metrics that matter come from the work, not from a conference slide. Deming’s call to institute training, drive out fear, and remove barriers implies deep respect for people. In practice, that means co-design metrics with the people who will use them. Ask operators how they know a day is going well, what early signs of trouble they trust, and which numbers are noisy. You will learn more in an hour on the line or in the clinic than in a week with spreadsheets.
In one distribution center, managers insisted that “picks per hour” was the right productivity metric. Floor workers shrugged. When we asked them how they judged a shift, the answers were different: congestion at aisles 3 and 5, the scanner lag after 3 p.m., and the rate of replenishment misses. We added two simple measures, aisle congestion minutes and replenishment miss rate, and watched picks per hour rise without pushing people to sprint. The workforce felt seen, and the numbers finally mapped to reality.
Working across Deming’s 14 principles to design a metric set
No single principle writes your dashboard. They work as a system. Here is a compact way to build a metric set using Deming’s lens, from aim to action:
- Start with constancy of purpose. Write the aim of the system in concrete terms including user, benefit, and constraints. From that aim, select a few outcome metrics that matter to the user, not just to the department. Pair each with drivers you can influence. Define each metric operationally. Establish decision rules grounded in variation. Visualize results over time, not as point-in-time league tables. Assign joint ownership across silos for measures that cross boundaries. Publish not just targets but the methods to reach them. Review metrics in a cadence that matches the process rhythm, weekly for flow, monthly for quality, quarterly for cost. Treat misses as learning opportunities, not hunting expeditions for culprits.
This second and final list compresses many of Deming’s principles into a practical sequence. It keeps the set small, the definitions clear, and the actions focused on the system rather than individuals.
Examples from three domains
To make this less abstract, consider how a Deming-informed metric design looks in three very different settings.
Software delivery. Many teams adopt the well-known DORA metrics, then get stuck chasing the numbers. A Deming approach starts with the aim, for instance, “safely deliver customer-visible improvements weekly with minimal disruption.” Outcomes become customer-impacting incident minutes per week and customer adoption of shipped features. Drivers include deployment frequency, lead time for change, change failure rate, and time to restore, but they are measured with control charts and tied to explicit methods like trunk-based development, automated tests, and peer review. The team avoids ranking individuals and looks instead for system constraints such as slow environments or flaky tests. Targets come with methods: if change failure rate exceeds a control limit, pause risky modules and invest in test fixtures. Leadership rewards early disclosure of risk, not hero pushes.
Healthcare ambulatory clinic. The aim might be “provide same-day access for urgent needs with high clinical quality and patient trust.” Outcomes include same-day appointment availability and return visit rates for the same complaint within 7 days. Drivers include schedule template adherence, triage response times, and guideline-concordant care for common conditions. The clinic treats variation seriously, comparing like-for-like days and locations, and resists naming and shaming clinicians for outlier weeks. Patient-reported measures are operationally defined, for example, trust measured via a validated short survey after visits, sampled to avoid survey fatigue. Cross-functional ownership is real: front desk, nursing, and clinicians share the same access and quality chart, and huddles focus on process adjustments, not personal admonitions.
Discrete manufacturing line. The aim could be “deliver finished units meeting spec at stable takt with minimum rework and safe working conditions.” Outcomes include first pass yield, on-time completion, and recordable incident rate. Drivers include station cycle time capability, andon pulls resolved within target, and supplier part conformance. The plant uses visual management with control charts at each station and a single system-level chart on throughput. Slogans are absent. Instead, you see operational definitions posted beside gauges and torques. Leadership resists the temptation to push overtime to hit end-of-week targets when control charts show common cause variation. They improve fixtures, adjust work balance, and collaborate with suppliers using shared measures. When a special cause occurs, the andon event log links to the downstream exception, creating feedback from effect to cause.
Across all three, the Deming mindset does the heavy lifting. The measures are familiar. The difference is in how they are chosen, defined, and used.
Avoiding common pitfalls that masquerade as rigor
Several traps recur when teams “get serious” about metrics.
Data purity at the expense of learning. Teams spend months perfecting a data warehouse before using a single measure to run an experiment. Better to start with credible, if imperfect, data and tighten definitions as you learn. Deming’s focus on prediction supports iterative refinement.
Overstuffed dashboards. A 40-card dashboard confuses. Choose a handful of measures tied to the aim, then rotate supporting charts into reviews when the narrative calls for them. If a chart never shapes a decision, retire it.

Vanity normalization. Percentages without denominators mislead. Always tie rates to counts and exposure. A 10 percent failure rate on 20 transactions is not the same story as 10 percent on 20,000.
Individual ranking. If you must look at individual variation, do it as a coaching conversation with privacy and context, not as a public scoreboard. Most signals at the individual level reflect system issues unless you have overwhelming evidence otherwise.
Targets without trade-offs. If you set an aggressive cycle time target, expect quality pressure. Make the trade-offs visible, and protect the metrics that reflect your nonnegotiables, such as safety and reliability.
Embedding the rhythm: review, prediction, and adjustment
Metrics work when they live inside a cadence. Deming’s plan-do-study-act is a natural home. In practice, I like a weekly operational review anchored by time-series charts and a monthly strategic review that examines whether the measures still match the aim.
In the weekly, teams bring annotated charts. An annotation ties a change in the process to a shift in the data. “New queue triage workflow deployed here,” or “Supplier X changed packaging.” Over time, the chart becomes a history of learning. The leader’s role is to ask predictive questions: given the last eight points above the median, what do you expect next week? What would disconfirm that expectation, and how will you respond? That habit builds intuition aligned with variation, not with wishful thinking.
In the monthly, leadership checks for drift. Has the aim changed? Do the current drivers still have leverage? Which measures caused action, and which sat flat? If a metric invites gaming, either fix the definition or replace it. This is not metric churn. It is stewardship.
Bringing Deming to the analytics stack
Modern analytics make it easy to slice, dice, and publish. The technology is not the hard part. Still, a few implementation details help:
Choose visualization that highlights variation and time. Control charts, run charts, and cumulative flow diagrams beat heat maps and traffic light grids for operational Visit this page measures. Resist point-in-time bar charts that hide volatility.
Keep operational definitions close to the chart. A small info icon with source systems, filters, and timestamp reduces confusion. If two teams use the same label with different logic, you have not yet done the hard work.
Build alerts on decision rules, not thresholds alone. If a special cause rule fires, notify the owner with context. Do not page people for random noise.
Tag events in your data model. Feature releases, supplier changes, staffing shifts, and policy updates should be first-class events that can be overlaid on charts. That one change turns many post-hoc debates into clear stories.
What Deming’s 14 principles change about your next metric
The phrase deming 14 principles can feel abstract until a number drives the wrong behavior or fails to drive any behavior at all. Here is what changes when you design with Deming in mind.
You start from purpose, so your metrics serve the user and the business, not the loudest voice. You respect variation, so you stop overreacting to noise and start investing in process capability. You define measures operationally, so debates shrink and improvement grows. You remove fear, so people surface problems while they can still be fixed. You stop counting inspection as progress and instead build systems that produce quality by design. And you pair targets with methods, so you measure what you can actually influence and accept the trade-offs openly.
That is the heart of metrics that matter. They teach you about the system, they point to levers you can pull, and they reinforce the culture you want. Done well, they make improvement almost inevitable. Not because the numbers are clever, but because they are honest about the work.