In general, the heuristics used by the statistical process control engineers are reasonable if you are using materials developed by Donald Wheeler, W. Edwards Deming, or the grandfather of these methods – Walter Shewhart.
But my question: Is there an error in the text for rule one that describes a shift? If I did the math right, a binary event with a probability of 0.5 that occurs exactly 6 consecutive times is \frac{1}{2^6} = 0.015625, for 7, it is \frac{1}{2^{7}} = 0.0078125 neither of which I consider close enough to “less than 3 in 1000”.
A long time ago, I wrote a program to explore the relationship among various measures of information described in the link below; the numbers in the quoted text looked wrong to me. Is there a more charitable interpretation, or did the authors simply get the computations wrong?
Related Threads
-
Thread with table relating powers of 2 with p values
Relating S-values to other measures of statistical information -
A discussion of reasonable uses of p values as decision rules in Statistical Process Control
Frequentist vs Bayesian debate applied to real-world data analysis: any philosophical insights? - #11 by R_cubed