The Knowledge System
The Knowledge System Podcast
Five-minute Deming: Plan-Do-Study-Act
0:00
-10:03

Five-minute Deming: Plan-Do-Study-Act

The four-step discipline behind real improvement.

Many management teams are praised for speed. They launch new initiatives and talk about momentum as if motion itself were evidence of progress. But fast action without disciplined learning creates a different problem: we spread assumptions through the system before we know whether they are sound.

That is why W. Edwards Deming’s Plan-Do-Study-Act matters. It gives leaders a way to slow down certainty without slowing down improvement. In the long run, it produces better service, lower waste, and a steadier reputation.

Why leaders need more than a pilot

Plan-Do-Study-Act (PDSA) is often described as an improvement cycle. That is true, but it can sound smaller than Deming intended. PDSA is a way to connect theory, prediction, action, and learning.

Plan means more than choosing an idea. It means stating what you think is happening, what change you want to test, and what you predict will follow. Do means carrying out that test, usually on a limited scale. Study means comparing the result with the prediction and taking surprises seriously. Act means deciding whether to adopt the change, abandon it, or run another cycle with a better theory.

Deming put the underlying point simply: “Management in any form is prediction.”

Management in any form is prediction.
— W. Edwards Deming

That is what many change efforts skip. We move from concern to action without ever being clear about the theory behind the action. Then we mistake activity for learning, or a short-term result for proof.

A story from commercial property management makes the problem easy to see.

What Harbor Point learned by slowing down

At Harbor Point Property Group, the executive team was under pressure. Tenants in three downtown office buildings were complaining about slow maintenance work, repeat visits, and weak communication from the service desk. Renewal season was approaching, and nobody wanted owners asking why routine service felt unreliable.

Claire, the head of operations, opened a Monday meeting with a familiar managerial move. She wanted speed, clarity, and a visible response.

“We need faster resolution times. I want every building manager under four hours for routine maintenance requests by next month.”

It sounded decisive. Complaints were rising. The pressure to look responsive was real.

But Jordan, the regional operations director, had spent the previous week reading work-order notes from the buildings. He saw something Claire’s demand did not explain. Some tickets stayed open too long. Others were closed quickly, then reopened. Vendor dispatches were inconsistent. Tenant descriptions were often incomplete. The pattern looked messy, not simple.

When Claire pressed him, Jordan answered with the line that changed the meeting.

“I think we know the symptom. I’m not sure we know the problem yet.”

That was the turning point. Instead of accepting a broad portfolio-wide push for faster close times, Jordan proposed a PDSA cycle. One building. One category of request. Two weeks. Plumbing calls in Franklin Tower only.

“Two weeks feels slow,” Claire said.

“Only if we confuse motion with learning,” Jordan replied.

This was the Plan stage, and he made it concrete. The service desk would ask three new intake questions before dispatching a plumber. Building staff would classify each request by severity. Vendors would receive tighter work orders with tenant access details and photos when available. Jordan’s prediction was clear: first-visit completion would improve, repeat visits would fall, and tenant updates would improve even if average close time did not improve right away.

That kind of planning is not paperwork.

It is disciplined thinking.

As Deming wrote: “Step 1 [Plan] is the foundation of the whole cycle.”

Step 1 [Plan] is the foundation of the whole cycle.
— W. Edwards Deming

The Do stage followed. For two weeks, Franklin Tower used the revised intake method only for plumbing calls. The service desk logged the new questions. Building staff tagged urgency consistently. Jordan reviewed requests daily to make sure the test was being carried out as planned.

Then came Study. The headline result was mixed. Average close time improved only slightly. If Harbor Point had judged the test by a single visible metric, the effort might have been dismissed as disappointing.

But the rest of the evidence told a more useful story. First-visit completion improved sharply. Repeat visits fell. Complaints about poor communication dropped. And one surprise stood out: the biggest delays were not coming from the plumbers. They were coming from incomplete tenant access information and late approvals for after-hours entry.

Claire saw it immediately. The dispatch script had helped, but not in the way they first expected.

“Right,” Jordan said. “We learned more than whether the average moved. We learned where the friction actually is.”

That answer captured the real value of the cycle.

That led to Act. Harbor Point kept the stronger intake questions, added a clearer path for access approvals, and ran another cycle in a second building with a different tenant mix. The second round confirmed some of the original theory and corrected the rest. The intake method held up. The approval issue mattered even more than they first thought. One vendor adapted quickly; another needed coaching.

In time, Harbor Point did standardize parts of the process. But they did not do what the first meeting had nearly produced. They did not issue a broad demand to close tickets faster and hope for the best. They acted on what the cycles taught them. Tenant complaints fell. Repeat work declined. Service became more reliable.

That is what PDSA looks like in practice.

Not delay. Learning strong enough to justify action.

Where we usually go wrong

Most of us do not resist PDSA because we dislike learning. We resist it because pressure makes immediacy feel responsible. When complaints rise, costs increase, or customers get restless, we want to show movement.

That impulse is understandable. It is also risky.

We often confuse a visible response with a thoughtful one. We roll out a policy, tighten a target, or announce a new standard before we have stated the theory behind it. Then, when numbers move, we read the movement as proof. If the result looks better, we congratulate ourselves too early. If it looks worse, we abandon the effort too quickly.

In both cases, we may learn very little.

Another problem is that we study outcomes too narrowly. We look for one summary number to tell the whole story. But systems rarely teach in a single measure. A useful test may reveal that the real issue lies in a handoff, an approval path, a vendor interaction, or a classification rule.

This weakens internal performance. Organizations that learn poorly create rework, inconsistency, and distrust. Organizations that learn well build steadier service and stronger trust.

What leaders can do instead

  1. Make prediction explicit. In every serious improvement effort, ask what theory is being tested and what result is being predicted. If we cannot answer those two questions, we are not ready to learn from the effort.

  2. Start smaller than your urgency prefers. A limited test is not a retreat from action. It improves the quality of action by creating learning with less cost and disruption.

  3. Study more than the headline result. A test may miss its most visible target and still reveal something crucial about approvals, timing, handoffs, demand, or coordination. The lesson may sit beside the metric we first cared about.

  4. Treat surprises as valuable evidence. When results differ from the prediction, resist the urge to defend the original idea. The gap between expectation and reality is often where the system becomes visible.

  5. Use Act as a leadership decision, not a ritual ending. Adopt what clearly helped. Abandon what did not. Revise the theory where the test exposed weak assumptions, then run another cycle. That is how management capability grows.

The discipline behind better improvement

The misconception is easy to understand: if a problem is urgent, the answer must be faster action. Deming’s point is different. Better results come from better learning, and better learning comes from method.

PDSA gives leaders that method. It asks us to think before acting, predict before judging, study before declaring success, and act only after the system has taught us something worth trusting.

That is not slower management. It is wiser management. Over time, it leads to what every organization says it wants: a better theory, a better system, and a better future built one thoughtful cycle at a time.

Without theory, there is no learning.
— W. Edwards Deming

Thanks for reading The Knowledge System! Subscribe for free to receive new posts and support my work.

Discussion about this episode

User's avatar

Ready for more?