Categories
Lean Six Sigma Uncategorized

Understanding Variables

In Six Sigma we talk a lot about X and Y variables.  We have to talk this way because, after all, Six Sigma is statistics.  Then you hear about ‘lead’ and ‘lag’ measures.  What are these and how do they relate to X and Y variables.  And what about ‘dependent’ and ‘independent’ variables.  Where do they fit in?  How do I know what to measure and when?  Relax.  It’s really not that complicated so lets step away and pull the covers back a little on some of the mystery.

X and Y and Independent and Dependent Variables

Ok…maybe you already knew this but these are the same thing.  X is the algebraic representation for the independent variable and Y represents the dependent variable.  Dependent or independent of what?  Well…of each other.  Let’s use a simple example to illustrate:

Let’s say it’s the New Year and you just made a resolution to loose weight (the most common of all resolutions and the one least likely to be kept by the way).  But your weight loss or gain doesn’t just happen in a vacuum. There are things that you decide to do to help you loose weight.  You work out, control your calories and your eating patterns for instance.  I have just described one dependent variable (Y) and three independent variables (X’s).

  • Dependent (Y) Variable – Your weight.  It is dependent on the other variables of how often and long you work out, how many calories you consume and whether you eat half a pint of Ben and Jerry’s Urban Bourbon ice cream at 11pm like I just did (I feel shameful but it was really gooood).
  • Independent (X) Variables – These are the variables that drive the Y variable.  The more you work out, the more you control your calories and the less you eat a large meal right before bed, the more you move the needle on the scale.

What is Y=f(X)

Do you suffer from algebraphobia?  Same here but this is really easy.  This equation is used commonly in Six Sigma to represent the relationship between the X and Y variables.  Read ‘Y=f(X)’ as ‘Y equals f of X’ or ‘Y is a function of X’.  Think of a function as a process.  X’s go into your function and a Y comes out.  The type, magnitude or other characteristic of X, determines the output of the function Y.  My weight loss (Y) is a function of my exercise habits (X1) or the extent to which I control my calories (X2) or how often I eat a big dinner late at night (X3).  Got it?  Cool.

Lead Versus Lag Measures

Ok…so we went from algebra to this lead and lag thing.  Again this is easy peasy.  ITS THE SAME THING!  Lead measures are the X values and lag measures are the Y values.  The reason we use lead and lag in this way is to describe their temporal relationship.  A lead measure is something you measure that is predictive of what will come after, or, the measurement that lags behind.  If I track my calories and keep them under a certain threshold everyday, I can reasonably expect the effect that lags behind will be weight loss.  The lag measure (weight loss) depends on the extent to which the lead measure (controlling calories daily) is met .

Think of lead measures as the lever that moves a big rock.  Consistent pressure on the lead measure, will eventually lead to the desired lag measure.  We act on and measure the lead measures as we are trying to affect change.  Measuring the lag measure doesn’t have any effect.  It only tells us at the end if we met our goal or not.  If I woke up everyday and all I did was weigh myself and say: “Boy, I’m really going to loose some weight today”, I probably would not be very successful.  But if I woke up every day and focused on the things in my control, my three X variables, it’s likely I’ll drop some lbs.

When to Measure What

In your improvement efforts, there are a lot of things that you will find that can be measured.  The key is to find the primary Y that you want to affect first and then find the levers to pull that are most likely to affect it…the X’s.  Once you have identified these key relationships, you first measure the Y (185lbs) before making any changes to the X’s.  This is your baseline measurement taken in the…yep…you guessed it…Measure phase of the DMAIC process.  Once you establish the baseline, then measure the X’s.

  • How many calories am I currently consuming daily? (3,000…yikes)
  • How often do I exercise each week? (2 times…er…ok 1 time)
  • How often do I eat after 9:00pm each week? (4 times…not including Ben & Jerry’s)

These are your X baselines; your levers for changing the Y.  Adjust them up or down to affect a change from the baseline in the target Y variable.  Then, measure the Y again to see the effect.

These concepts are cornerstones of process improvement and quality management principles.  Try to think of some other examples in every day life of lead and lag measures, X and Y variables and instances where Y=f(X).  Once you start to play with these concepts a bit they get easier and easier to see.  You begin to see causal relationships as levers to be pushed and pulled to deliver an outcome.

Categories
Uncategorized

You Can’t Understand Traffic Flow by Just Counting Cars

If you counted the number of cars on a stretch of road every morning at the same time, and found that the number of cars was the same every day, would you draw the conclusion that none of the cars are moving?  Probably not.  Yet that is exactly what we do when we make inferences about the health of a process by only looking at the process’ inventory over time.  To understand how well a process is functioning, we must consider all three legs of the process measurement stool; Average Inventory, Throughput Rate and Cost.  The goal is always to increase throughput while decreasing inventory and cost.

I often see people make the mistake of looking at month over month counts of process inventory (e.g. number of in process work orders, expense reports, recruiting orders, sales orders etc.) and drawing conclusion from the change in count.  To carry forward the traffic analogy, in order to understand the flow efficiency, we need to know how many cars are entering and exiting the stretch of road and an average time between entry and exit.  These additional measures paint a picture of the flow of traffic (throughput) not just a static snap shot.

To really understand the efficiency of this stretch of road, there is one more measure to consider; Cost.  Capturing the cost to maintain the road now gets at the heart of the matter; the cost to move a car from one end of the road to the other.

With these three measures now in place, we can watch this stretch of road for variation in performance.  Significant changes in any of these three measures triggers research to determine cause.  Consider this scenario:

Number of cars on the road (inventory) decreases, and number of cars exiting the road (throughput) decreases and the maintenance costs increase – While a decrease in inventory is good, it comes at the expense of decreased throughput and increased cost.  After investigating the cause of these changes, we find that significant roadwork is being performed to build an additional lane.  This work closed one of the three existing lanes.  The lane closure reduced the capacity of the road meaning fewer cars can be on the road at any one time.  The reduced capacity also meant that cars were closer together and drivers were reducing their speed which compounded to result in a significant reduction in the number of cars exiting the road.  The work being performed was increasing the current costs, and, because there would now be another lane to maintain, increased the future maintenance costs as well.

Can you think of similar lane closures in your business processes?  What about when a key team member resigns?  Or the disruption caused by a change to a new tool or process.  What about organizational changes?  These are examples of temporary lane closures that impact business process health.  Armed with the right measures of Average Inventory, Throughput Rate and Cost, you have an early warning system that will alert you to problems sooner and improve your ability to assess the impact of changes on the overall health of your processes.

See also Three Process Metrics that Matter and Theory of Constraints

 

Categories
Uncategorized

A Little Process Math

In an earlier post, I talked about the Three Process Metrics that Matter; process inventory, throughput and cost.  These three metrics are the cornerstone to understanding how a particular process is performing. But, sometimes it can be difficult to find the average inventory of a system. Taking periodic sampling is one way to do it. Simply count the units in the process at several intervals and take the average to get average inventory. There are a few problems with this method however. For one thing, it is time consuming. It can also lead to inaccuracies caused by natural ebbs and flows of a process. For instance, say you decide to take the sampling of a trouble ticketing system every morning. Because the night shift is lightly staffed, the system naturally has a higher volume of tickets than later in the day after the fully staffed shifts have worked down the list. In this case, your average inventory will be artificially high. You could randomize your sampling schedule to take samples at different times throughout the day, but this can start to get rather complex and laborious. A better way is to use Little’s Law to calculate the average inventory over time.

In 1954, John Little published a paper on queuing theory in which he postulated that the average number of units in a stable system is equal to the average arrival rate of new units multiplied by the average time to process a single unit. This is the algebraic expression of Little’s Law:

Or, putting it a bit more simply:

Inventory = Throughput * Time

While the original description refers to “arrival” rate (as in, the arrival of customers to a store), the formula works equally as well to use the “exit” rate (as in, the rate at which sales orders are completed). Therefore, if we know for instance that we completed 50 sales orders in 10 days, our rate is 5 orders per day (throughput). And, assume we know that the average amount of time to process a single order is 2 days, then we can solve for the average inventory in the system.

Inventory (L units) = Throughput (5 orders per day) * Time (2 days)

Average Inventory = 10

Rearrange the formula to find any one of the missing three variables.

Time = Inventory / Throughput

OR

Throughput = Inventory / Time

A couple points of caution when using Little’s Law. First, be sure to keep your units of measure the same. If you are measuring throughput in days and process time in hours, you will need to convert hours to a portion of a day (or vice versa). The other thing to be aware of is Little’s Law does not account for units already present in a system at the start of the testing interval or units that have not yet completed when the testing interval is stopped. Using a longer testing interval tends to decrease the mathematical effects of these two conditions. If you need to keep your testing intervals short, you may need to find a way to estimate the leading and lagging units.

Little’s Law can save a great deal of time and produce more consistent results when measuring process performance. Try it out next time your measuring a process and amaze your friends and coworkers.

Categories
Uncategorized

Three Process Metrics That Matter

Measuring process performance is the cornerstone of continual improvement. This may seem simple and obvious but like so many things, while it may be simple, it is often not easy. Sometimes it is difficult to capture the data necessary to maintain valuable measurements but I find this is not usually the problem. The more common problem is measuring too much of the wrong things. The technology that supports and enables business processes provide almost any data we want. Like Sherlock Holmes we think “Data, data, data; I can not make bricks without clay.” But, unlike fiction, more data often creates doubt and obscurity instead of leading us straight to the villain. There are only a critical three metrics that matter when gauging the health of any process. All other metrics are valuable as further forensics only after a problem is detected.

So, why not measure anything we can? I mean, we capture all this data, isn’t it a waste if we don’t use it? Consider Segal’s law; “a person with 2 watches is never sure of the time”.   Segal’s law is a cautionary postulation against the pitfalls of using too much information in decision making. If a person is wearing two watches, there are three potential states of those watches:

  1. Both watches are accurate and showing the same time. In this case, the wearer is confident because there is validation between instrumentation.
  2. One or both watches are inaccurate and displaying different times. In this case, the wearer is doubtful of the correct time because instrumentation is conflicting.
  3. Both watches are displaying the same inaccurate time. In this case, the wearer is just as confident of the correct time as if both watches are working putting trust in validated inaccuracy. Uh oh, then, if both watches are telling the same time, then that casts doubt on condition one.

So there is never a case where the wearer of 2 watches is truly confident of the correct time. Then the focus becomes on keeping instrumentation in synch. One well maintained high quality metric is far more useful than any combination of lower quality and often conflicting measurements.

The Theory of Constraints tells us there are only three key metrics to worry about when measuring the health of any given process; inventory, throughput and cost.

Inventory is all of the units in the process at any given time. Units are what the process acts on to produce an output. Some examples of business process units might include a sales order, trouble ticket, invoice, expense report or change order as example.

Throughput is the rate (expressed as # units over time period) at which a process produces output that is usable by all downstream processes. The last part of this is critical because it answers the common question “What about quality metrics?”. Quality is embedded in the throughput metric because only outputs that meet quality standards for all downstream processes are counted as throughput. If you apply this rule ruthlessly, it is not uncommon to find processes that, when first measured, have a throughput of zero because every output requires some level of rework, or does not meet service level’s or contains defects that are corrected by a downstream process.

Costs include all the money used to produce process throughput. Actual costs of a single process can be very difficult to capture. Most of the time, you’re looking at a single process acted on part time by many people for instance. Don’t get too wrapped up in this. If you can’t get to actual dollars, find a viable surrogate. In business processes, this is usually some variation of a human capital measurement.

When used together, these three metrics paint a valuable picture of the health of a single process. The objective is to reduce inventory and control or reduce costs while increasing throughput. Statistically relevant variations in the patterns of these three measures direct us to specific problems and areas requiring attention. Additionally, since each process is measured in a similar way, we can now make better decisions where to focus our improvement efforts. Given the choice between working on only one of two different processes, we can now compare relative changes in inventory, throughput and cost. Assuming both processes are valid candidates for improvement efforts, the one with the greatest deviation should get the most focus (all other things being equal).

Think of process measurement like a trip to the doctor. The first measures are simple indicators of health like weight, temperature and blood pressure. The results of these basic measures may drive further analysis but you’d never get wheeled in for a CAT scan or a full panel blood work up as a first step.

Don’t get seduced by the “more is better” mantra when it comes to process measurement. Reports that show a wall of data are rarely effective at driving action. Look for the key measures that are the best representation of inventory, throughput and cost and use those as your bellwether for action. Additional data may be required to select and direct specific actions. Use supporting data for forensic purposes not to measure top level process health. Start simple and use deeper level data to analyze once a problem is detected.