Featured Video
This Week in Quality Digest Live
Metrology Features
James Rawstron
Jet wing surface scan reduced to under four minutes
Michael Huda
How do you control a color that absorbs so much light, spectrophotometers can’t measure it?
Matthew Martin
Traditional CMMs are being challenged by fast-emerging, highly advanced blue-light scanning
Dirk Dusharme @ Quality Digest
Christmas is creeping, should a four-year degree really be the minimum requirement, and what we learn from Disneyland
Dean Solberg
If you can think about it and model it, you can 3D-print it

More Features

Metrology News
Four times faster data acquisition, enabling precise measurement at the submicron level
Caters to the growing need for sophisticated scanning in an entry-level solution that is budget friendly
Topics for the Go Forth and Measure project are virtually unlimited
Ask questions, exchange ideas and best practices, share product tips, discuss challenges in quality improvement initiatives

More News

Donald J. Wheeler

Metrology

Is the Part in Spec?

Since data is always imperfect, how can we ever know if a measured item is within the specifications?

Published: Tuesday, June 1, 2010 - 09:02

During the past 20 years it has become fashionable to condemn measurement processes that are less than perfect. Yet the reality is that we must always use imperfect data. Given this fact of life, how can we ever know if a measured item is or is not within the specifications? Put another way, how can we make allowance for measurement error when characterizing product relative to specifications?

To complicate the answer to these questions, we have measurements that have measurement units attached, and we have the uncertainty in those measurements that are expressed in terms of measurement units squared. Much of what has been written about these problems has been flawed. However, rigorous treatments of these issues are available, and in this article I shall present the results of them.

The uncertainty in a measurement

The uncertainty in any measurement may be defined by the standard deviation of repeated measurements of the same item. In practice this quantity may be estimated by having one operator measure a set of items two or three times each, using the same instrument, as follows:

Product 42LF is made in small batches. The small batch size, along with the vigorous stirring that is part of the process, makes it reasonable to consider each batch as effectively homogenized. A single test sample is drawn from the batch while it is being stirred. This test sample is split in half and the viscosity is measured twice. The test values, in centistokes, for seven batches are shown below:


Batch

78

79

80

81

82

83

84

Viscosities

2480

2870

2350

2990

3070

3020

2510

2530

2730

2390

2930

3050

3000

2610

Ranges

50

140

40

60

20

20

100

The duplicate measurements provide a simple and effective way of assessing the measurement error. The ranges shown above represent test-retest error of the same product, and therefore may be used to compute an estimate for the standard deviation of test-retest error:

Est. Std. Dev. for Test-Retest Error = Average Range/d2 = 61.4/1.128 = 54.4 Centistokes

When your boss asks what that means, what are you going to say? The value of 54.4 centistokes is an estimate of the square root of the rotational inertia about the averages for the set of test-retest values. Do you want to offer this as an explanation to your boss? If not, then you need a value that is easier to explain in terms of the problem, and in 1818 Wilhelm Bessel came up with just such a value: the “probable error.” The probable error is simply 0.675 times the standard deviation of test-retest error. It defines the median amount by which a measurement will err: Half the time measurements of a known standard will differ from the accepted value by more than one probable error, and half the time these measurements will differ from the accepted value by less than one probable error.

Thus, the probable error defines the effective resolution of a measurement. If a measurement will differ from the “best value” by an amount greater than the probable error at least half of the time, then there will not be much point in attempting to interpret the number more precisely than plus or minus one probable error.

The probable error defines the essential uncertainty in the measurement. As long as the probable error is larger than the measurement increment, it will define the effective resolution of a single value. Thus, the probable error defines the effective discreteness of a measurement, and it serves as a guide for how closely we should interpret a single value. For the example above, our estimate for the probable error of the viscosity measurements would be:

Probable Error = 0.675 (54.4 cs.) = 36.7 Centistokes

The individual measurements of viscosity are recorded to the nearest 10 centistokes, but they are good to the nearest 37 centistokes. Half the time they will err by more than 37 centistokes, and half the time they will err by less than 37 centistokes.

As soon as we know the probable error, we also have an answer to the age-old question of how many digits to record. The measurement increment should be approximately the same size as the probable error:

Smallest Effective Measurement Increment = 0.2 Probable Error

Largest Effective Measurement Increment = 2 Probable Errors

When the measurement increment falls outside this range you will be recording too many or too few digits. (A rigorous explanation of this result is provided in Chapter 13 of my book EMP IIIEvaluating the Measurement Process and Using Imperfect Data (SPC Press, 2006). For the viscosity measurements, the smallest effective measurement increment would be:

0.2 Probable Error = 0.2 (37 cs.) = 7 Centistokes

Thus, recording the viscosities to the nearest 10 centistokes is appropriate. Recording to the nearest centistoke would be excessive because the last digit recorded would be pure noise.

Likewise, rounding these measurements off to the nearest 100 centistokes would also be a mistake because the largest effective measurement increment here is 73 centistokes. Rounding to the nearest 100 centistokes would involve throwing away useful information.

Watershed specifications

Since the probable error defines the effective resolution of a measurement, it will also define the effective increment to use in fuzzing up the specifications to make allowances for measurement error. However, before we begin to adjust our specifications, we need to make allowance for the fact that specifications are generally stated in terms of acceptable values. That is, if I tell you the specifications are six to 12, you will understand that a value of six will be acceptable, as will a value of 12. If the measurement increment is whole numbers, the first unacceptable values will be five and 13. Because our computations are going to work with numbers as if they come from a continuum, we must make a continuity correction and define the “watershed specifications” as:

Lower Watershed Specification = Minimum Acceptable Value – One-half Measurement Increment

Upper Watershed Specification = Maximum Acceptable Value + One-half Measurement Increment

Thus, when our specifications are six to 12, and parts are measured to the nearest whole number, our watershed specifications would be 5.5 to 12.5. With these values our specified tolerance would now be computed as:

Specified Tolerance = 12.5 – 5.5 = 7 Units

And our computation would now match the reality that there are seven acceptable values within the specifications of six to 12.

When our measurement increment is small relative to the specified tolerance, this adjustment will become trivial and can be ignored. However, as the measurement increment gets large relative to the specified tolerance, this adjustment becomes essential for good computations.

Fuzzing the specifications

The problem of using imperfect data is the problem of how to define manufacturing specifications such that when a measurement falls within the manufacturing specifications, the product is likely to be within the customer specifications. In figure 1 we see that there are two key aspects to this problem, the distance D and the definition of “likely.”

Figure 1: The idea behind manufacturing specifications

 

Since the probable error defines the effective resolution of a measurement, it also defines the natural unit to use in fuzzing the specifications to allow for measurement error. In Chapter 14 of EMP III, I go through the whole mathematical argument and obtain the following results.

Figure 2: 64-percent manufacturing specifications = watershed specifications

If you use the watershed specifications as your manufacturing specifications, you will have items that will fall within one probable error of one of the watershed specifications at least part of the time (unless your measurement increment is too large). Although you can say that these items have at least a 64-percent chance of being in spec, the fact that some of the items you are shipping have this small chance of conformity means that the most you can say about your product stream is that it has at least a 64-percent chance of being conforming product. Thus, when you use the watershed specifications, you are effectively using 64-percent manufacturing specifications.

Figure 3: 85-percent manufacturing specifications

On the other hand, if you tighten the watershed specifications by one probable error on each end, then all the product you will end up shipping will have at least an 85-percent chance of conforming. Thus, 85-percent manufacturing specifications are the watershed specifications tightened by one probable error.

Figure 4: 96-percent manufacturing specifications

If you tighten the watershed specifications by two probable errors on each end, then all of the product you will end up shipping will have at least a 96-percent chance of conforming. Thus, 96-percent manufacturing specifications are the watershed specifications tightened by two probable errors.

Figure 5: 99-percent manufacturing specifications

If you tighten the watershed specifications by three probable errors on each end, then all of the product you will end up shipping will have at least a 99-percent chance of conforming. Thus, 99-percent manufacturing specifications are the watershed specifications tightened by three probable errors.

Figure 6: 99.9-percent manufacturing specifications

If you tighten the watershed specifications by four probable errors on each end, then all of the product you will end up shipping will have at least a 99.9-percent chance of conforming. Thus, 99.9-percent manufacturing specifications are the watershed specifications tightened by four probable errors.

The likelihoods cited above are a posteriori probabilities computed under the assumption that an unmeasured item had a 50-50 chance of conformity (hopefully you are doing better than this). This is why the likelihoods are given as minimums.

Given the trade-off between tighter specifications and higher likelihoods, and given that the likelihoods are minimums, I recommend using 96-percent manufacturing specifications in most situations.

Thus, once you have some knowledge about the test-retest error of your measurement system, you can compute the probable error and use this value to determine the correct number of digits to record and to obtain manufacturing specifications that answer the perennial question, “How can we be sure that we are shipping conforming product?”

Discuss

About The Author

Donald J. Wheeler’s picture

Donald J. Wheeler

Dr. Donald J. Wheeler is a Fellow of both the American Statistical Association and the American Society for Quality, and is the recipient of the 2010 Deming Medal. As the author of 25 books and hundreds of articles, he is one of the leading authorities on statistical process control and applied data analysis. Find out more about Dr. Wheeler’s books at www.spcpress.com.

Dr. Wheeler welcomes your questions. You can contact him at djwheeler@spcpress.com

Comments

Is the part in spec?

As a tier 1 supplier to the auto industry, I have to insure that any product shipped to my customers (OEM's) meets their specifications. I also have to insure that my suppliers have capable processes and also ship parts to me that meet the specification. When problems occur (and they alsways do!), we are called upon to react quickly to protect the OEM customer. Often this includes temporary 100% inspection while corrective actions are determined and implemented. While costly, this is much cheaper than shutting down a customer assembly plant or incurring a product recall.

If you do not take the measurement uncertainty into account when sorting products from a process that is not capable, then the risk is great that you will accept non-conforming product. This article is spot on, and the actions it recommends should be considered mandatory for containment plans involving sorting parts. It is at our company.

A nice summary

This is a nice summary of the topics covered in Depth in Wheeler's book. Perhaps a comparison of this technique and the GUM would be a good topic for future columns.

Parts in Spec and Measurement Uncertainty

I was hoping to find an article that simplified the confusing world of measurement uncertainty. Instead we now have Dr. Wheeler's uncertainty calculation in addition to the system defined by the ISO Guide to Uncertainty of Measurement (the GUM). While I like Dr. Wheeler's method because it is simpler, I cannot ignore the GUM and its calculations.

I think it would have been better if Dr. Wheeler had mentioned that his system was different than the system that governs calibration and test laboratories that are accredited to ISO 17025.

I hope that the next few years will bring more clarity to this important subject.

Jack Dearing