Our PROMISE: Our ads will never cover up content.
Our children thank you.
Published: Tuesday, June 1, 2010 - 10:02
During the past 20 years it has become fashionable to condemn measurement processes that are less than perfect. Yet the reality is that we must always use imperfect data. Given this fact of life, how can we ever know if a measured item is or is not within the specifications? Put another way, how can we make allowance for measurement error when characterizing product relative to specifications?
To complicate the answer to these questions, we have measurements that have measurement units attached, and we have the uncertainty in those measurements that are expressed in terms of measurement units squared. Much of what has been written about these problems has been flawed. However, rigorous treatments of these issues are available, and in this article I shall present the results of them.
The uncertainty in any measurement may be defined by the standard deviation of repeated measurements of the same item. In practice this quantity may be estimated by having one operator measure a set of items two or three times each, using the same instrument, as follows:
Product 42LF is made in small batches. The small batch size, along with the vigorous stirring that is part of the process, makes it reasonable to consider each batch as effectively homogenized. A single test sample is drawn from the batch while it is being stirred. This test sample is split in half and the viscosity is measured twice. The test values, in centistokes, for seven batches are shown below:
Batch |
|||||||
Viscosities |
|||||||
The duplicate measurements provide a simple and effective way of assessing the measurement error. The ranges shown above represent test-retest error of the same product, and therefore may be used to compute an estimate for the standard deviation of test-retest error:
Est. Std. Dev. for Test-Retest Error = Average Range/d_{2} = 61.4/1.128 = 54.4 Centistokes
When your boss asks what that means, what are you going to say? The value of 54.4 centistokes is an estimate of the square root of the rotational inertia about the averages for the set of test-retest values. Do you want to offer this as an explanation to your boss? If not, then you need a value that is easier to explain in terms of the problem, and in 1818 Wilhelm Bessel came up with just such a value: the “probable error.” The probable error is simply 0.675 times the standard deviation of test-retest error. It defines the median amount by which a measurement will err: Half the time measurements of a known standard will differ from the accepted value by more than one probable error, and half the time these measurements will differ from the accepted value by less than one probable error.
Thus, the probable error defines the effective resolution of a measurement. If a measurement will differ from the “best value” by an amount greater than the probable error at least half of the time, then there will not be much point in attempting to interpret the number more precisely than plus or minus one probable error.
The probable error defines the essential uncertainty in the measurement. As long as the probable error is larger than the measurement increment, it will define the effective resolution of a single value. Thus, the probable error defines the effective discreteness of a measurement, and it serves as a guide for how closely we should interpret a single value. For the example above, our estimate for the probable error of the viscosity measurements would be:
Probable Error = 0.675 (54.4 cs.) = 36.7 Centistokes
The individual measurements of viscosity are recorded to the nearest 10 centistokes, but they are good to the nearest 37 centistokes. Half the time they will err by more than 37 centistokes, and half the time they will err by less than 37 centistokes.
As soon as we know the probable error, we also have an answer to the age-old question of how many digits to record. The measurement increment should be approximately the same size as the probable error:
Smallest Effective Measurement Increment = 0.2 Probable Error
Largest Effective Measurement Increment = 2 Probable Errors
When the measurement increment falls outside this range you will be recording too many or too few digits. (A rigorous explanation of this result is provided in Chapter 13 of my book EMP III—Evaluating the Measurement Process and Using Imperfect Data (SPC Press, 2006). For the viscosity measurements, the smallest effective measurement increment would be:
0.2 Probable Error = 0.2 (37 cs.) = 7 Centistokes
Thus, recording the viscosities to the nearest 10 centistokes is appropriate. Recording to the nearest centistoke would be excessive because the last digit recorded would be pure noise.
Likewise, rounding these measurements off to the nearest 100 centistokes would also be a mistake because the largest effective measurement increment here is 73 centistokes. Rounding to the nearest 100 centistokes would involve throwing away useful information.
Since the probable error defines the effective resolution of a measurement, it will also define the effective increment to use in fuzzing up the specifications to make allowances for measurement error. However, before we begin to adjust our specifications, we need to make allowance for the fact that specifications are generally stated in terms of acceptable values. That is, if I tell you the specifications are six to 12, you will understand that a value of six will be acceptable, as will a value of 12. If the measurement increment is whole numbers, the first unacceptable values will be five and 13. Because our computations are going to work with numbers as if they come from a continuum, we must make a continuity correction and define the “watershed specifications” as:
Lower Watershed Specification = Minimum Acceptable Value – One-half Measurement Increment
Upper Watershed Specification = Maximum Acceptable Value + One-half Measurement Increment
Thus, when our specifications are six to 12, and parts are measured to the nearest whole number, our watershed specifications would be 5.5 to 12.5. With these values our specified tolerance would now be computed as:
Specified Tolerance = 12.5 – 5.5 = 7 Units
And our computation would now match the reality that there are seven acceptable values within the specifications of six to 12.
When our measurement increment is small relative to the specified tolerance, this adjustment will become trivial and can be ignored. However, as the measurement increment gets large relative to the specified tolerance, this adjustment becomes essential for good computations.
The problem of using imperfect data is the problem of how to define manufacturing specifications such that when a measurement falls within the manufacturing specifications, the product is likely to be within the customer specifications. In figure 1 we see that there are two key aspects to this problem, the distance D and the definition of “likely.”
Since the probable error defines the effective resolution of a measurement, it also defines the natural unit to use in fuzzing the specifications to allow for measurement error. In Chapter 14 of EMP III, I go through the whole mathematical argument and obtain the following results.
If you use the watershed specifications as your manufacturing specifications, you will have items that will fall within one probable error of one of the watershed specifications at least part of the time (unless your measurement increment is too large). Although you can say that these items have at least a 64-percent chance of being in spec, the fact that some of the items you are shipping have this small chance of conformity means that the most you can say about your product stream is that it has at least a 64-percent chance of being conforming product. Thus, when you use the watershed specifications, you are effectively using 64-percent manufacturing specifications.
On the other hand, if you tighten the watershed specifications by one probable error on each end, then all the product you will end up shipping will have at least an 85-percent chance of conforming. Thus, 85-percent manufacturing specifications are the watershed specifications tightened by one probable error.
If you tighten the watershed specifications by two probable errors on each end, then all of the product you will end up shipping will have at least a 96-percent chance of conforming. Thus, 96-percent manufacturing specifications are the watershed specifications tightened by two probable errors.
If you tighten the watershed specifications by three probable errors on each end, then all of the product you will end up shipping will have at least a 99-percent chance of conforming. Thus, 99-percent manufacturing specifications are the watershed specifications tightened by three probable errors.
If you tighten the watershed specifications by four probable errors on each end, then all of the product you will end up shipping will have at least a 99.9-percent chance of conforming. Thus, 99.9-percent manufacturing specifications are the watershed specifications tightened by four probable errors.
The likelihoods cited above are a posteriori probabilities computed under the assumption that an unmeasured item had a 50-50 chance of conformity (hopefully you are doing better than this). This is why the likelihoods are given as minimums.
Given the trade-off between tighter specifications and higher likelihoods, and given that the likelihoods are minimums, I recommend using 96-percent manufacturing specifications in most situations.
Thus, once you have some knowledge about the test-retest error of your measurement system, you can compute the probable error and use this value to determine the correct number of digits to record and to obtain manufacturing specifications that answer the perennial question, “How can we be sure that we are shipping conforming product?”
Comments
What to do with products rejected for tight tolerance limits?
I asked Dr. D. J. WHEELER for further guidance on products that were rejected within tight tolerance limits.
Dr. DONALD J. WHEELER kindly agreed to give me an explanation. For which I am very grateful to him.
Here are the links provided to me to explain further actions of the article:
1. Where Do Manufacturing Specifications Come From?
2. 100% Inspection and Measurement Error
Respectfully
Sergey Grigoryev
Is the part in spec?
As a tier 1 supplier to the auto industry, I have to insure that any product shipped to my customers (OEM's) meets their specifications. I also have to insure that my suppliers have capable processes and also ship parts to me that meet the specification. When problems occur (and they alsways do!), we are called upon to react quickly to protect the OEM customer. Often this includes temporary 100% inspection while corrective actions are determined and implemented. While costly, this is much cheaper than shutting down a customer assembly plant or incurring a product recall.
If you do not take the measurement uncertainty into account when sorting products from a process that is not capable, then the risk is great that you will accept non-conforming product. This article is spot on, and the actions it recommends should be considered mandatory for containment plans involving sorting parts. It is at our company.
A nice summary
This is a nice summary of the topics covered in Depth in Wheeler's book. Perhaps a comparison of this technique and the GUM would be a good topic for future columns.
Parts in Spec and Measurement Uncertainty
I was hoping to find an article that simplified the confusing world of measurement uncertainty. Instead we now have Dr. Wheeler's uncertainty calculation in addition to the system defined by the ISO Guide to Uncertainty of Measurement (the GUM). While I like Dr. Wheeler's method because it is simpler, I cannot ignore the GUM and its calculations.
I think it would have been better if Dr. Wheeler had mentioned that his system was different than the system that governs calibration and test laboratories that are accredited to ISO 17025.
I hope that the next few years will bring more clarity to this important subject.
Jack Dearing