Cost for QD employees to rent an apartment in Chico, CA. $1,200/month. Please turn off your ad blocker in Quality Digest

Our landlords thank you.

Statistics

Published: Wednesday, April 1, 2015 - 15:40

The first axiom of data analysis is: “No data have meaning apart from their context.” Yet we all encounter measures that have been dreamed up without regard for either context or use. This column gives a couple of the more egregious examples I have encountered over the years.

The first example is a formula used by the Kentucky Higher Education Commission in the 1970s to allocate state funds to the various state-run colleges and universities. A former student of mine sent this formula along with the relevant data for all of the post-secondary state schools. This funding equation had five terms that had to be evaluated for each school. The first four terms were relatively complex terms involving various variables to characterize the school, while the last term was the simple ratio of the school’s full-time equivalent head count to the total full-time equivalent head count for the state as a whole.

Because of the complexity of the first four terms I began by simply evaluating the formula for each school. Since I wanted to understand the formula, I evaluated each term separately. I listed these in a spreadsheet with the colleges down the left side, and the values for each of the five terms in the successive columns. As I did this I found that the first and third terms were always large and positive, while the second and fourth terms were always large and negative. Of course, in each case, the fifth term was a simple proportion.

As I worked with these values I began to notice a curious phenomenon. For each school the sum of the first four terms always came out to be zero! The first four terms had nothing to do with funding. The funds were prorated on the full-time equivalent head count, and the remainder of the equation was there simply to confuse the school administrators and to keep the legislators at bay.

In the 1990s the U. S. Postal Service used the Delivery Collection Efficiency Achievement (DCEA) measure to ensure that “the beatings will continue until quality improves.” The *DCEA* measure was a function of the four variables that I will call *X*, *Y*, *Z*, and *H*:

*X* = Total cased delivery volume for the operating unit. If all the mail processed by the unit in a single month was placed in a single stack, this value would be the height of that stack measured in feet.

*Y* = Delivery days in the month

*Z* = Possible deliveries. Different units serve different numbers of addresses. So the product of *Y* and *Z* will essentially define the total number of addresses served by the unit in a month.

*H* = Man-hours for the operating unit

Clearly, the volume *X*, divided by the man-hours *H*, would be a simple measure of productivity, but the *DCEA* is slightly more complex than this. In fact it begins with the ratio:

If we interpret *X* to represent the delivery volume of the operating unit, and if we think of the product of *Y* and *Z* as the area of opportunity for deliveries, then perhaps it might make sense to divide *X* (production volume) by the product of *Y* and *Z* (number of addresses served) to get a measure of efficiency [volume per address]. However, when this ratio is raised to the 0.35922 power all vestiges of rationality are lost. Both the ratio and the attached measurement units become complete nonsense when raised to the 0.35922 power. How could you even begin to explain this value? Next the *DCEA* multiplies the nonsense ratio above by the following sum:

The first term of this sum is a very, very specific fraction (given down to the level of parts per million) of our “area of opportunity,” the product of *Y* and *Z*. The second term in this sum is a simple measure of productivity—volume divided by man-hours. The problem here is that there is no way that these two terms will have the same units. The first term will be a count of [total delivery addresses served] while the second will be a ratio of [feet per hour]. If you ignore the units attached to your values, then you too can end up adding apples and zebras to get gobbledygook.

Finally, the *DCEA* multiplies the product of the nonsense ratio and the gobbledygook sum by the obvious scaling factor of 5.61439! Any questions?

Thus, the *DCEA* is nothing more than numerical jabberwocky—complex nonsense intended to impress the numerically illiterate. Such meaningless numbers may be computed, but they cannot be called *measures* since they do not, in any real sense, measure anything. If you can’t articulate exactly what a value measures or counts, then you have left the realm of common sense.

Just because you can compute a piece of numerical jabberwocky does not justify its use. The only justification for using any number is that it makes sense in terms of its context. Statistical techniques cannot undo numerical jabberwocky. No matter how sophisticated your analysis, when you fail to respect the primary axiom of data analysis and end up using numerical jabberwocky in your computations, your results will be triumphs of computation over common sense.

As we saw in the last example, one of the easiest ways to create numerical jabberwocky is to take some measure, which makes sense on its own, and transform that measure in some nonlinear fashion. Logarithmic transformations, exponential transformations, square-root transformations, and trigonometric transformations can quickly transport you to a land of fantasy and make-believe. Once there, statistical techniques cannot restore that which was lost.

So, “Beware the Jabberwock, my son!”

## Comments

## Nice

A lovely, subtle attack on the Six Sigma crazies and their meaningless transformations. Why do people want to buy Minitab and spend their hours playing with nonsense numbers instead of reading Shewhart?

## Brillig ADB

There is light enough for those who wish to see, and darkenss enought for those who are otherwise inclined.