A report of how a process performs is not only a function of process characteristics and sampling chance differences. It can also depend on sampling approach. For example, one person could describe a process as out of control, which would lead to activities that address process perturbations as abnormalities; another person could describe the same process as being in control.
ADVERTISEMENT |
To illustrate how different interpretations can occur, let's evaluate the time series data in figure 11, which could be the completion time for five, randomly selected daily procedural transactions in a hospital, insurance company, or one-shift manufacturing facility. The data initially record process stability and then, if stable, the process's capability relative to customer specifications of 95 to 105.
…
Comments
Three-Way Chart
Don't know enough about this process to say for sure if it is appropriate, but another alternative is to use a three-way chart as described in Donald Wheeler's book "Understanding Statistical Process Control".
Individuals vs. Xbar charts
Agree with the previous post, that 3 way charts (within/between) are a good way to monitor two sources of variation (e.g. within cavity, between cavity).
Also regardless of the presence of other sources of variation between subgroups, individuals charts will always be less sensitive to detecting process changes than will be xbar charts. Smaller shifts are detectible as the sample size increases. The inherent potential Type II errors found in Individuals charts seesm to be rarely discussed (assuming we are using SPC to detect small process changes when they occur).
Rational Subgrouping and Sampling
Whenever I see an Average/Range chart that looks like the one presented in this article, I always question the subgroup rational. What souces of variation are within and between subgroups? Is the Xbar/R chart the most appropriate one based on context and the data collection scheme?
I would start off with the XmR chart as a first step in your analysis.
Rich DeRoeck
Wrong use of probability plot and transformation
I agree with the fundamental argument that if the variability between subgroups is significantly larger than the variability within subgroups, your X-Bar chart will have limits that says everything is out of control. Limits for the averages should be developed via an ANOVA method (or ANOM – essentially treating the averages as individuals) until such time (if ever) that they discover the source as to why the variability between groups is so large. If one does a quick ANOVA, you find that 96% of the total variability is the variability between “days”.
However, there are two big flaws in the analysis here. Where this went off the track is the notion that one uses a probability plot with all 50 data points (appropriate if you have rational subgroups which is not the case). The graph and subsequent analysis implies that the points are all from the same distribution and clearly they are not. In a situation such as this, where the between group variability overwhelmingly dominates the within variability, a sample size of 1/day is more than sufficient. The conclusion this probability plot provides that the data is not normal is erroneous and is further exacerbated by implying a transformation is required. If one does an Anderson-Darling statistic on the daily averages (n=10), the p-value is 0.27, therefore not rejecting the hypothesis that the data is normal. Even if a transformation is required (which it isn’t), blindly accepting a recommendation of a lognormal distribution makes absolutely no physical sense. Lognormal distributions are inherently used when the data spans several magnitudes and that’s not the case here.
Response to previous comments
Thanks for your comments:
1. Three-way control charting is applied, for example, when there is variability within a part and several parts are selected within subgroups that are tracked over time using a control chart. This was not the case in the example described; however, the three-way control chart does use an individuals chart of subgroup means, which is similar to the described 30,000-foot-level reporting.
2. Agree that the hypothesis of normality from the probability plot would be rejected for this data set; hence, the estimates of non-compliance will have some associated error because of this. However, even for this extreme example, where there are two distinct distributions, the probably plot does not look as bad as one might initially think from being a straight line; hence, we can still provide a rough estimate of the percentage not conforming. In all likelihood, this non-normality topic would not have been brought up if the capability statement were presented in the article using Minitab’s capability analysis routine, which does not include a normality assessment when providing its process capability reporting. The probability plot non-conformance result noted in the article is the same as what could have been reported using Minitab’s capability analysis routine output "PPM total". Also, the 30,000-foot-level charting approach would lead to the appropriate behavior – understanding what could be done to improve the process.
3. Agreed that the next step that one would take when attempting to reduce the amount of non-conformance in the 30,000-foot-level charting would be to create hypotheses to test out various theories. One theory might create a hypothesis test for assessing variability between days relative to within days, which would be significant. This could then lead to investigation why this is occurring and what might be done to reduce the source of this between subgroup variability. If an improvement were made, the 30,000-foot-level control chart would transition to a new level of stability that had a reduced rate of non-conformance shown in the probability plot. Note how this behavior is quite different than addressing all the special cause conditions that were created using traditional control charts, which could have led to much firefighting.
4. A control charting of standard deviation could have been made without a transformation. The standard deviation would have been in control for this set of data; however, zero would have been within the control limits. Since standard deviation cannot physically be below zero, this leads us to a dilemma. We could have moved the control limit on the lower side to zero; however, there would have been no way to detect if the within-subgroup variability reduced because of a change. A more general approximate approach to address this situation is to take the log of standard deviation. Transformations should be made only when they make physical sense. Since standard deviation can never get below zero, the log-normal distribution is a general, easy to use, transformation that fits fairly well and makes physical sense for this situation.
5. Subgrouping is very important. With 30,000-foot-level reporting, we need to have all common-cause input variability to occur between subgroups.
6. Relative to the type II error comment, the primary goal of a 30,000-foot-level assessment is to not detect small changes but describe at a high level how the process is performing relative to stability and customer needs. When our common-cause variability from this high-level view is not providing what we desire then process improvements are needed. Improvements to the process are then demonstrated when the 30,000-foot-level chart transitions to a new, improved level of stability. Note, this is different than traditional control charting which has a primary intent of identifying when special cause occurs so that these problems can be addressed in a timely fashion.
Transformation confusion followup
#2 - “Rough estimate of the percentage not conforming” – here is where Wheeler makes a strong argument that we should not transform. Our estimate is just that. It is an estimate that has variability. With the data that Breyfogle presented, I bristle at the mere mention of transformation. Where does that obsession come from? It is not borne in the data. There is no justification for it and if we use voodoo statistics, we do become our own worst enemy. Even Breyfogle agrees that it is not from the same population – therefore we NEVER transform just to transform so some black-box statistic comes out. If we want to live in theory and not in reality, then the process capability is 0, because the process is not in control. Obviously managers will not tolerate such a thing. So why are we trying to be so precise if it is “out of control”? If you accept the argument that it should be the subgroup averages that should be tracked (which I think we both agree – I would conclude and agree with any who state there are really 10 data points to worry about – the subgroup averages, given that over 96% of the variability is between), then why are we going through distributional gyrations with 50 data points? I would agree with the paper if it had stopped at the 3 way graph and had gone no further.
As for the control charting of the standard deviation in #4, color me confused. You can easily do a s-chart vs a R-chart. For this, the lower limit is zero for n=5 according to my constants table. Why make it harder? Maybe I didn’t understand your point and I’ll await a reply. But we know from statistical theory that the distribution of standard deviations for low sample sizes is not normally distributed. Why would we force it to be when we already have established methods to deal with it?
Response to: Transformation confusion followup
Sorry about not responding sooner, but I did not see this comment until someone recently point it out to me.
It is interesting that even mentioning the word data transformation and this can draw strong emotions from some people. One needs to keep in mind that Wheeler, when making the statement that transformations are not necessary, is using a traditional control charting strategy where focus is given to control charting individual processes; e.g. creating a control chart for each machine individually, where ten machines may be manufacturing a part. Wheeler's primary emphasis is not making process capability statements of how the overall process (e.g., from 10 machines) is doing relative to specification requirements.
What I described in this article is not real-time control charting of individual processes but 30,000-foot-level performance tracking over time, with the inclusion of how the overall process is doing relative to specification needs; i.e., collectively evaluating all ten machines over time and providing a process capability/performance statement relative to how well the overall process (from the 10 machines) is performing relative to customer needs.
When one is making a process capability statement, it is important to keep in mind that an appropriate transformation be considered otherwise a good fit cannot be made for the estimate; i.e., like what is needed for making any estimate in engineering from a model. Some data are not normally distributed by their nature. For example, the time it takes to complete a process. The lower boundary is zero. If the process operates near this boundary, then there the distribution will tend to have a skewed distribution. A log-normal distribution can fits this situation well as a model for making a process capability/performance statement.
Responses to the specific points made in the comment:
* From a 30,000-foot-level approach, the process was in control; hence, a process capability/performance statement could be made. This statement is even predictive, which means that if this non-conformance rate is undesirable the process needs to be improved.
* The 3-way graph provides no process capability/performance statement, which the 30,000-foot-level charting provides; hence, the 3-way control chart is not adequate for what this article suggests.
* With the R-bar chart providing a 0 boundary, it can be very difficult to determine if the within subgroup variability of the process improved, since it is physically impossible to have an out of control condition below zero. This shortcoming is overcome with 30,000-foot-level chart of standard deviation that has a log-normal transformation.
control charting & capabilitying
I wonder wether USA is an acronym for Using Statistical A ...: if Statistix were an effective Tool, its so much intensive use should have resulted in a better World, whatever the meaning of this Statement. Which is not; for many reasons; may be the root cause - or reason - is that Statisitix is like the mythical Panacea - just a myth. Thank you.
Add new comment