Dec 10 2017
Is SPC Obsolete? (Revisited)
Six years ago, one of the first posts in this blog — Is SPC Obsolete? — started a spirited discussion with 122 comments. Reflecting on it, however, I find that the participants, including myself, missed the mark in many ways:
- My own post and comments were too long on what is wrong with SPC, as taught to this day, and too short on alternatives. Here, I am attempting to remedy this by presenting two techniques, induction trees and naive Bayes, that I think should be taught as part of anything reasonably called statistical process control. I conclude with what I think are the cultural reasons why they are ignored.
- The discussions were too narrowly focused on control charts. While the Wikipedia article on SPC is only about control charts, other authors, like Douglas Montgomery or Jack B. Revelle, see it as including other tools, such as scatterplots, Pareto charts, and histograms, topics that none of the discussion participants said anything about. Even among control charts, there was undue emphasis on just one kind, the XmR chart, that Don Wheeler thinks is all you need to understand variation.
- Many of the contributors resorted to the argument of authority, saying that an approach must be right because of who said so, as opposed to what it says. With all due respect to Shewhart, Deming, and Juran, we are not going to solve today’s quality problems by parsing their words. If they were still around, perhaps they would chime in and exhort quality professionals to apply their own judgment instead.
Feb 27 2019
Analysis of Oscar TV Viewership | Mark Graban | LinkedIn
“Initial analysis says Oscars TV viewers and ratings are up over 2018. But, for context, the ratings were DOWN in 2018 compared to 2017.
Instead of reacting to each new data point, look at trends over time using “run charts” or “process behavior charts” (as I created). Did having “no host” make any statistically significant difference in ratings? No. The numbers are just fluctuating. Don’t waste time cooking up an explanation for ‘noise’ in a metric.”
Source: LinkedIn
Michel Baudin‘s comments: What happened between 2004 and 2005? Visually, the downward step is the one feature of the chart that stands out. If the technique Mark used is valid, there should be assignable causes from that period to explain the drop. If you forget the green and red lines and just look at the time series, however, a smooth, long-term downward trend sounds like a better fit. Perhaps Mark should try more modern trend analysis tools besides the venerable XmR chart.
Following a day of spirited discussion, Mark sent me the data and asked for my own analysis, which is in the chart below. The gray band is a 99% confidence interval, calculated without any knowledge of the nature of the numbers. The points outside the band can be assumed to have assignable causes for their unusual high or low ratings, and I took a look at the press reports on a few. As you can see, there is nothing special happening in 2004-2005.
The popularity of a show has nothing to do with a critical dimension on a manufactured part, which is supposed to be constant. The plant organization has the authority to do what it takes to keep it that way. There is no reason why show ratings should be constant.
In addition, while a length measurement qualifies as a raw data point, the published number of viewers of a TV show doesn’t. In the US, it is an estimate based on a poll of 5,000 households. Viewership of the Oscars ceremony in 2019 was 11.5% higher than in 2018. On 5,000 data points, it does not look like a fluctuation. By comparison, political polls use samples of 1,000 to 1,500 voters and claim margins of error of ±3%.
The viewership estimates aren’t just numbers in a table, and you can’t interpret them properly without knowing their backstory.
#SPC, #XmRChart, #ProcessBehaviorChart
Share this:
Like this:
By Michel Baudin • Blog clippings 2 • Tags: Process Behavior Chart, SPC, XmR Chart