Open access peer-reviewed chapter

Shewhart Control Chart: Long-Term Data Analysis Tool with High Development Capacity

Written By

Vladimir Shper, Elena Khunuzidi, Svetlana Sheremetyeva and Vladimir Smelov

Submitted: 27 June 2023 Reviewed: 24 November 2023 Published: 23 December 2023

DOI: 10.5772/intechopen.113991

From the Edited Volume

Quality Control and Quality Assurance - Techniques and Applications

Edited by Sayyad Zahid Qamar and Nasr Al-Hinai

Chapter metrics overview

68 Chapter Downloads

View Full Metrics

Abstract

This chapter suggests some of the ways in which we can enrich our understanding of the theory of variability when we extend our attention to a gap between the real problems any practitioner may encounter and the traditional theory of control charts stated in textbooks, guides, standards, etc. The benefits are about more than just covering additional ground, for this expanded focus also provides insights into how many real problems are being ignored, many new types of charts turn out to be excessively difficult for engineers, many tacit assumptions that traditional theory is based on stay not being understood by practitioners. We are going to consider the impact of different types of process instability, data homogeneity, nonnormality, and nonrandomness on the right application of Shewhart control charts. We also propose the recommendations to practitioners on how to avoid the above-mentioned problems and improve data-based decision-making.

Keywords

  • exploratory data analysis
  • shewhart control chart
  • process stability
  • assignable causes of variation
  • nonhomogeneity
  • nonrandomness
  • capability indices

Useful application of control charts starts and ends with education

Kaoru Ishikawa

Advertisement

1. Introduction

On 16 May 2024, the World will be celebrating the 100th anniversary of the Shewhart Control Chart (ShCC) – the most essential tool of process stability analysis used successfully in practically all areas of human activity. This 100-year-old tool of exploratory data analysis is described in many old and new books (see, e.g., [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11], to name a few), and there are international standards devoted to control charting [12] as well as numerous sites on the Internet. On the other hand, most quality professionals, managers and engineers, CEOs, and even not a small amount of statisticians either are not familiar with this tool of data analysis or have just very superficial understanding of it [13]. Almost 30 years ago, R. Hoyer and W. Ellis claimed this thought in their paper ([14], 63)1 in the following form: “… our experience indicated that a sizable majority of quality professionals are not knowledgeable about basic issues of statistics and Statistical Process Control (SPC). Our instructional activities in a broad range of academic, industrial, and service delivery environments have convinced us that there are many individuals who are “doing SPC” without understanding what it is about”. Then they made a conclusion (ibid): “Although it is disappointing that the technical content of the quality improvement discipline has progressed so little during the past 65 years, that is probably not the most significant problem facing process control initiatives during the next decade. Instead, there is every reason to be concerned that many quality professionals are directing continuous process improvement activities without a sound understanding of the basic issues”. We are sure that this situation has not changed notably since those times [13], but this chapter is devoted just to “the technical content of the quality improvement discipline”. It is noteworthy that such prominent experts in Statistical Process Control (SPC) area as Lloyd Nelson and William Woodall argued emphatically [15] against the assertion of Hoyer and Ellis that “the technology of the quality science has been intellectually dormant for the past 65 years” ([16], 73). The authors of [14] agreed with the experts that many interesting issues have been going on in the research journals devoted to quality improvement. “But, unfortunately, it appears to be primarily academics… and there are many real-world settings in which these results are only marginally useful” ([15], 93). That is just what our paper is focused on. We will discuss several traditional assumptions that cover a minuscule part of reality and lead to many gaps between real process behavior and corresponding math models. Moreover, there are many practical questions that are not being discussed in the current literature at all, and have not been ever discussed in the past. In Section 2, we will give a brief survey of the problems that seem to be the most important to us. Then, Section 3 is devoted to the ambiguity in the notion of assignable causes of variation – one of the basic ideas of SPC. In Section 4, we will present our considerations on the data nonrandomness. Then, in Section 5, we will discuss the issues of highly skewed processes. Section 6 contains some examples revealing the limitation of current theory of ShCCs. Finally, in Section 7, we will share our concerns about the interaction between SPC and metrology, SPC and management. Our proposals for further research are given in conclusion.

Advertisement

2. The most important problems of Shewhart control chart application

It was outlined in [13] that “There is no friendship between business and the theory of variation – incomprehension is going on and on” [13]. This is not just our viewpoint. We have already mentioned above similar statements of Hoyer and Ellis in [14, 15, 16]. A well-known expert in SPC W. Woodall in 2000 published a survey, “Controversies and Contradictions in Statistical Process Control” [17]. One of the main problems discussed there was the relationship between hypothesis testing and control charting. The main Woodall’s conclusion on this issue is that it is a simplification to consider control charting as something equivalent to hypothesis testing ([17], 343). This approach can be a serious obstacle to the right application of ShCCs in Phase I of process analysis. We agree. Moreover, this widely spread view of equivalency between these two entities can prevent the right use of control charts both in Phase I and in Phase II. What is even more important is that very few practitioners know about this problem and hardly ever ponder on it.

D. Steinberg wrote in 2016 a survey of the state-of-the-art in industrial statistics, where he mentioned the following problems in SPC: multivariate data, profile data, and data from phasor measurement units [18]. He expressed his concern that too many research papers rely on unrealistic assumptions that do not exist in practice.

In 2017, W. Woodall wrote a follow-up to his 2000 survey called “Bridging the gap between theory and practice in basic statistical process monitoring” [19]. This time, he revisited some of the same problems he had talked about before, and discussed some new ones. Among the old problems, there was again the relation between statistical theory and practice. At the end of that paper, Woodall made several useful suggestions that could have improved the quality of statistical papers in the area of SPC and the quality of related studies. Simultaneously, he made a proposal that we think is completely unacceptable. According to Woodall, the use of the moving range chart should be ceased. We beg to differ from this suggestion and will explain our viewpoint below.

In 2022 W. Woodall published a new paper on SPC issues titled “Recent Critiques of Statistical Process Monitoring Approaches”. In the introductory section, he wrote: “Hundreds of flawed papers on statistical process monitoring (SPM) methods have appeared in the literature over the past five to ten years. The presence of so many flawed methods, and so many incorrect theories, reflects badly on the SPM research field. Critiques of the various misguided approaches have been published in the last two years in an effort to stem this tide” [20]. Let us look at the flawed methods enlisted by Woodall: Use of Inadvisable Weighted Averages, Use of Auxiliary Information, Rules Equivalent to Runs Rules, Neutrosophic Methods, Mixing Various Charts, The Generally Weighted Moving Average Chart, Misuses of the EWMA Statistic, Repetitive Sampling Methods, using the coefficient of variation, the multivariate coefficient of variation, and various capability indices, etc.

It is noteworthy that overwhelming majority of practitioners uses only simple ShCCs because many new types of charts (e.g., CUSUM, EWMA, changepoint, etc.) turn out to be too difficult for engineers, operators, and workers.

We are sure that there are at least two root causes of such a sad situation. One, mentioned above, was described in Hoyer, Ellis paper [14, 15, 16]. Another one is more fundamental. In the report written in 1996, G. Box noted [21]: “An important issue in the 1930s was whether statistics was to be treated as a branch of Science or Mathematics. Unfortunately, to my mind, the latter view has been adopted in the United States and in many other countries. Statistics has for some time been categorized as one of Mathematical Sciences, and this view has dominated “university teaching, research, the awarding of advanced degrees, promotion, tenure of faculty, the distribution of grants by funding agencies and the characteristics of statistical journals”. Judging by above-mentioned papers nothing has changed since 1996. All flawed techniques enlisted by Woodall in [20] are math’s exercises or “Statistical Gymnastics” as caustically noted Ch. Quesenberry in [22]. Shewhart’s close friend and associate W. Edwards Deming, ending the foreword to the 1939 Shewhart’s book, wrote: “Another half-century may pass before the full spectrum of Dr. Shewhart’s contributions has been revealed in liberal education, science, and industry” ([23], ii). It seems like another half-century may pass before all who are trying to use control charts efficiently have understood the main ideas of Shewhart and Deming.

D. Steinberg, in his paper [18], cited well-known statistician B. Gunter, who wrote in 2008 panel discussion in Technometrics on the future of industrial statistics: “I fear that Technometrics has evolved from primarily making connections to the real, hard, and complex questions of scientific practice to primarily producing artificial formulations of those questions suitable for compact “solution” by mathematical characterization. To understand what is useful and not merely wrong in industrial statistical practice, we need to pay much more attention to the messy details that make up reality”. Then Steinberg did not agree with Gunter’s statement that most academic papers “has become completely cut off from real problems”. But he agreed “that many of the most challenging and exciting problems arising today are not getting space in our journals and that we need better theory to guide us in attacking such problems” ([18], 52).

Let us sum up the main idea of all papers cited above: too many statistical works went far away from real practice and do not help practitioners in solving their real problems. This is a direct contradiction to Shewhart-Deming approach and to the basic idea of Shewhart control chart, which “stands out as the only one that actually examines the data for the internal consistency which is a prerequisite for any extrapolation into the future. Thus, unlike all “tests” and “interval estimates” of statistical inference Shewhart’s process behavior charts are tools for Analytic Studies. Rather than mathematical modeling, or estimation, Shewhart’s charts are concerned with taking appropriate actions in the future based upon an analysis of the data from the past. Out of all the statistical procedures available today, they alone were designed for the inductive inferences of the real world” ([9], 19). We see this tendency to disregard reality for the world of math models and also ignore the problems of simple control charts in favor of more and more complex designs. Many books and standards that are being widely used by practitioners all over the world teach the theory of control charts based on very unrealistic assumptions about real processes and their behavior (see [1, 2, 3, 4, 5, 7, 12], to name a few). In the following discussion, we will examine more thoroughly some issues, such as the various types of assignable causes of variations, the examples of unanswered questions in the theory of control charts, and other related topics.

Advertisement

3. What is an assignable cause of variation, and how it changes a process?

According to Shewhart ([24], 14), “…in the majority of cases there are unknown causes of variability in the quality of a product which do not belong to a constant system…these causes were called Assignable”. He, further, explains that an assignable cause of variation is one which can be found without excessive waste of time and money. What is mostly important for us, Shewhart outlines the principal impossibility to establish a criterion of revealing an assignable cause a priori either by formal or by mathematical method.

A famous statistician and quality guru, Dr. Deming wrote in his foreword to Shewhart’s book ([23], ii): “The great contribution of control charts is to separate variation into two sources: (1) the system itself (‘chance causes’, Dr. Shewhart called them), the responsibility of management; and (2) assignable causes, called by Deming ‘special causes’, specific to some ephemeral event …” A process is called statistically controllable or stable or predictable if all assignable causes of variation are removed.

On the other hand, Wheeler and Chambers [6] defined assignable causes of variation as follows: “Uncontrolled variation that is characterized by a pattern of variation that changes over time”. In a paper written later, Wheeler [25] pointed out that chance or common causes differ from assignable causes due to their impact on a process. So, they are not principally contradictory.

Woodall [17] provided the following definition: “‘Common cause’ variation is considered to be due to the inherent nature of the process and cannot be altered without changing the process itself. ‘Assignable (or special) causes’ of variation are unusual shocks or other disruptions to the process, the causes of which can and should be removed”.

Finally, Montgomery [10] stated that: “…common causes are sources of variability that are embedded in the system or the process itself, while assignable causes usually arise from an external source.”

Thus, there are slightly different views on whether assignable causes of variation are a result of intervention into the system from the outside or not; however, there is a full agreement that they are “some ephemeral events that can usually be discovered …and removed”. A tool for distinguishing assignable causes of variation from common causes is the control chart, coined by W. Shewhart in 1924. Up to 2010 there were more than 4000 research papers published on this topic [26]. Intensive analysis of books and main reviews in this area showed [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 25, 27, 28] that practically all papers on the ShCCs considered very simple model of chart’s behavior. Almost all researchers studied the statistical properties of simple ShCC charts when some assignable cause of variation changed either the mean or the standard deviation or both of the underlying distribution, which stayed of unchanged type (and almost always was normal).

Alternatively, if one looks through the explanation of assignable causes of variation in the very popular SPC Manual [29], used in auto industry for many years, she/he will see the picture (page 30 in [29]), which clearly shows that an assignable cause of variation can lead to an arbitrary change of the distribution function type2. Just this was the main idea of the work [30]. The authors studied the case‘when after a special cause of variation emerged, the underlying normal distribution function (DF) transformed into either uniform or log-normal distribution. It was found that the probabilities of detecting a shift in the mean changed radically from the case of normal DF (see Figures 4–7 in Ref. [30]). Indeed, as soon as one accepts the opportunity of DF type to change after the impact of assignable cause of variation, a lot of different possibilities emerges, and only one of them has been investigated and described in the literature. It was proposed in Ref. [30] to introduce two types of assignable causes of variation: not changing the underlying DF and changing it. It is worth stressing that this idea can be generalized, as ShCCs do not need any assumptions about DF type. So, a more general proposal could be an introduction of two types of assignable causes of variation: not changing the system where the process is going on, and changing that system. Though more than 10 years have passed since the paper [30] was published, this idea has not been either supported or refuted by statistical community. Here, we would like to revisit this notion from a different angle. Let us look at Figure 1, taken from our work [31]. One can see in Figure 1, the ShCC with two red circles and two green ovals on it. Red circles relate to the points falling beyond chart’s limits, and ovals show the points where the process mean jumped. Obviously, both situations emerged due to some special causes of variations. But is there any difference between these two cases: one when the assignable cause was evanescent, and the system has not changed, and second when the assignable cause has changed the system? As far as we know, such a question has never been discussed in SPC literature. Does it deserve to be discussed? We are sure it does because in the first case the search for the root cause of interference in the process has to be made by the process team (engineers, operators, linear managers, etc.), and in the second case this search is an act of top management – only CEOs are responsible for the system as a whole. This forces us to come back to the idea of different types of assignable causes of variations. Having generalized and slightly modified the definitions from [30, 31] we suggest the following version:

Figure 1.

Daily sales of a distribution network. Here 1 centner = 100 kilograms.

Definition 1: An assignable cause of variation of type I (Intrinsic) does not change the system within which a process works (e.g., does not change the type of the underlying DF). As a result, it is quite natural to consider such a type of assignable causes as belonging to the system (though this is not a necessary condition).

Definition 2: An assignable cause of variation of type X (eXtrinsic) changes the system within which a process works (e.g., changes the parameters, or type or both of the underlying DF). As a result, it is quite natural to consider such a type of assignable cause as, most probably, not belonging to the system (though this is also not necessary).

If the statistical community agrees with our suggestions, then the difference between dissimilar types of assignable causes of variation will help practitioners to grasp who, in the first place, has to interfere in the process. This is highly important knowledge to improve the process with success. The instability due to assignable causes of type I requires searching for a root cause within the system. The instability due to assignable causes of type X requires searching for a root cause outside the system.

Advertisement

4. The importance of time order of process values

As stressed by Shper and Adler [32], the problem of data nonrandomness has been underestimated in recent years, though it was of primary importance to Shewhart. More than once, W. Shewhart returned to this issue, explaining the key role of the order of points for understanding whether a process is stable or not. On page 12 of [23], he clarifies that all attempts to determine some DF that could thoroughly describe a state of statistical control are useless and senseless. Some statisticians considered the normal law to be such DF, but these hopes turned out to be refuted completely. Further on page 27, W. Shewhart continues: “…the significance of observed order is independent of the frequency distribution…” and “… are primitive”. Shewhart’s conclusion about the importance of the point order was firmly supported by such outstanding gurus as W. Deming (see Deming’s Foreword in [23]) and G. Box [33].

Let us consider a revealing example of an erroneous conclusion caused by the neglecting of the role of the order of points in real processes. A well-known expert in SPC, W. Woodall has long since his survey of 2000 [17], stood up for the elimination of the moving range chart from an arsenal of SPC tools. His arguments were based on the results of random data simulation, which follows from the paper of Rigdon et al. [34], which Woodall referred to in [17]. However, the moving range automatically takes into account the order of points due to the structure of successive differences. That is why it contains much more important information about the process than the standard deviation (SD). The SD completely ignores the succession of process points. Therefore, using the average of the moving range of two (AMR) to estimate the ShCC limits allows anyone to take into account the order of points within a process. If we eliminate the moving range chart from using, we automatically neglect many patterns within a process.

Another famous SPC expert, L. Nelson, also recommended avoiding using the moving ranges because of problems with their interpretation [34]. But, in fact, he considered the moving ranges necessary to calculate the limits of the x-chart because they are better than SDs. The reason is clear: the moving range measures variations from point to point irrespective of their level, which can vary due to trends, oscillations, patterns, etc. Now, one may ask: how often are any patterns present in real processes, and if they influence the outcomes or not?

Without any doubt patterns are present at all real processes, but sometimes their influence may be small, and, consequently, ignored. The famous Box adage about models is working, of course. But, in practice, the level of pattern influence is rarely known beforehand. Shewhart foresaw this many years ago: “… a sequence is called random if it is known to have been produced by a random operation, but is assumed to be nonrandom if occurring in experience and not known to have been given by a random operation” ([23], 17). This means that if an observed sequence is or is not random, it can be verified only in the future and not by any ingenious math. It follows immediately from this Shewhart remark that there is no and cannot even exist any universal indicator of process nonrandomness. So, what should we do in such a situation? We need to have a variety of dissimilar metrics/indices/rules revealing different types of nonrandomness. There are a number of such indices that are well-known and widely used. The so-called additional rules for ShCC interpretation are the first coming to mind. These rules are described in practically all books, guides, standards, etc., of SPC (see, e.g., §5.7 in [9]). It is noteworthy that each such rule can reveal only one single case of nonrandomness. In other words, the standard set of rules covers a minuscule part of potential opportunities. Except for these rules, there is a run test on randomness [35] – a useful rule based on the number of series in data. Again, it reveals only one type of nonrandomness – a relationship between a number of points lying above and below a chart central line.

A new test on data randomness was proposed in [32]. It is the ratio of AMR to SD. As noted above, the value of AMR is very sensitive to any patterns of nonrandomness, so this ratio deviates from its standard value as soon as data have some kind of nonrandomness3. What is important about this index is that it covers not a single case of nonrandomness but, rather, an unknown set of possibilities. In Figure 2 (taken from [32]), one can see the dependence of confidence limits for AMR/SD on the number of points (k) under investigation. If the value of AMR/SD lies above or below the lines on that picture, then one may state that with 95% confidence, the data are not random. The exact values of probabilities are given in [32].

Figure 2.

The limits of AMR/SD ratio for k = 10, 20, 30, 40, 50, 100, 150, and 200.

There was the question raised in Shper and Adler [32]: why do current studies of ShCCs ignore the order of points? Our version is as follows. This ignorance is caused by the lack of understanding of the difference between analytic and enumerative studies. As Deming explained many times (see, e.g., [37]), an enumerative study deals with units taken from some population that have a definite value of mean, SD, DF, etc. An analytic study deals with a process that is going on and on; it does not have a definite mean or DF because it changes permanently. Many statistical methods such as confidence intervals and hypothesis testing can be applied to enumerative studies but are inapplicable to analytical ones. Deming’s conclusion in Ref. [37] turned out to be quite unambiguous: “Statistical theory for analytic problems has not yet been developed” [37].

Advertisement

5. The impact of non-normality on the control limits of ShCCs

It is well-known that the traditional theory of ShCCs is based on (1) assumption of data randomness and (2) assumption of data being distributed normally. We discussed the first assumption in the previous paragraph. Below, we will consider the second.

It is assumed here that the reader is somewhat familiar with the construction and use of ShCCs. According to a generally accepted view, each ShCC has limits, separating the zone of system variability from the area of the assignable cause habitation. These limits can be calculated with very simple formulas based on the three-sigma rule suggested by Shewhart [24]. Typical formula for the control limits (CL) of many popular control charts for variables looks like [38]:

CL=Average±Scaling FactorSome Measure of DispersionE1

Scaling factors in (1) are frequently named as control chart constants and are usually denoted by different capital letters with indices, for example, A2, D3, D4, and E2 (to name a few most widely used). They, in turn, depend on the so-called bias correction factors d2, d3, and d4 ([9], 416):

A2=3d2nE2
D3=13d3d2E3
D4=1+3d3d2E4
E2=3d2E5
E5=3d4E6

The factors d2, d3, and d4 (and, therefore, chart coefficients) are considered constant in the SPC literature, mostly widely used by practitioners (e.g., [6, 8, 9, 10]). Though, statisticians working in the field have known for a long time that non-normality has a great impact on bias correction factors. Why is this? On the one hand, this opinion is based on the works of many outstanding statisticians of the first half of the last century (see references and other details in [39], and, e.g., §7.3 in the excellent book [40]). On the other hand, the bias correction factors do change insignificantly in many cases but not in all possible ones. This issue was carefully studied in our work [39], where such skewed distributions as exponential, log-normal, Weibull, Burr and Pareto were simulated, and the values of d2, d3, and d4 were estimated. The main results of that paper are as follows.

In conclusion to his landmark work of 1967 [41], Irwing Burr wrote: “… we can use the ordinary normal curve control charts constants unless the population is markedly non-normal. When it is, the tables provide guidance on what constants to use.” Unfortunately, Burr did not point out what the words “markedly non-normal population” meant operationally. Moreover, there has been no discussion on this issue up to now. So, we proposed in [39] that a twofold increase in the probability of falling beyond the control limits be considered as the condition of significant deviation. Then, after simulation, the results shown in Tables 13 were obtained. In Table 1, the parameters of investigated DFs and their notation are presented. β1 and β2 are the squared skewness and traditional kurtosis, respectively. The means of d2, d3, d4, and the corresponding values of A2, E2, and E5, calculated by using Eqs. (2)(6), as well as their relative deviations from the standard values for the x-mR chart, are given in Table 2. Table 3 provides similar information for X¯-R chart with subgroup size n = 2 and 3. The DFs having a relative probability increase of less than twofold are excluded from Table 3. Almost all figures in Table 2 show the values that relate to DFs having more than a twofold increase in corresponding probabilities. It was recommended in [39] that for all cases when one encountered “markedly non-normal” data, he/she should use the algorithm to construct the ShCC proposed in [39] and the corrected values of the chart’s constants given there.

#DFParameters of DFβ1β2Notation
1NormalMean = 0 SD = 10.003.00Gau
2UniformMean = 0 SD = 10.001.80Uni
3Exponentialλ = 0.1; 0.2; 0.014.009.00Exp
4Log-normalMean = 0 SD = 0.53.068.90Lgau1
5Mean = 0 SD = 0.65.1113.27Lgau2
6Mean = 0 SD = 0.78.3420.79Lgau3
7Mean = 0 SD = 1.038.25113.94Lgau4
8WeibullShape parameter = 0.87.9215.7Wei1
9Shape parameter = 1.51.154.40Wei2
10Shape parameter = 0.712.2423.54Wei3
11LogisticMean = 0 SD = 10.004.20Lgi
12Burr’sc = 2, k = 4 from Burr’s table I2.047.36B5
13ParetoShape parameter = 205.489.13Prt

Table 1.

Notations and parameters of DFs studied in [39].

d2d3d4E2E5D4δE2, %**δE5,%δD4, %
Gau*1.1290.8510.9552.663.143.26−0.10−0.12−0.21
1.1280.8530.9542.6603.1453.268
Uni1.1560.8161.0172.602.953.12−2.44−6.21−4.60
Exp1.0030.9940.6992.994.293.9712.4436.4721.58
Wei10.9271.0680.5673.245.294.4621.6668.2436.36
Wei21.0910.8990.8712.753.443.473.379.526.24
Wei30.8661.1120.4763.466.304.8530.23100.4048.48
Lgau11.0410.9570.7822.883.843.768.3421.9814.99
Lgau21.0030.9950.7182.994.183.9812.4432.8521.67
Lgau30.9651.0320.6553.114.584.2116.8745.6328.77
Lgau40.8271.1440.4563.636.585.1536.37109.1957.59
Lgi1.1040.8840.9022.723.333.402.165.754.11
B51.0670.9260.8352.813.593.605.7014.2410.27
Prt0.9781.0210.6583.074.564.1315.3244.9726.44

Table 2.

Results for x-mR chart.

This row gives the generally accepted values of bias correction factors.


δE2, % denotes the relative deviation of E2 from its standard value 2.66. Similar – δE5,% and δD4, %.


d2d3d4A2D4δA2,%**δD4, %
Gau*1.128
1.128
0.852
0.8525
0.953
0.954
1.881
1.880
3.266
3.268
0.03−0.06
n = 2
Exp1.0000.9980.6922.1213.99412.8422.22
Wei10.9211.0740.5602.3034.49822.5137.65
Wei30.8651.1200.4752.4524.88430.4549.46
Lgau11.1490.9950.9041.8463.598−1.8010.10
Lgau21.0050.9940.7182.1113.96712.2721.39
Lgau30.9561.0410.6422.2194.26718.0330.56
Lgau40.8181.1550.4472.5935.23637.9460.22
Prt0.9781.0270,6572.1694.15015.3727.00
n = 3
Gau*1.690
1.693
0.890
0.8884
1.584
1.588
1.025
1.023
2.578
2.575
0.180.12
Exp1.5041.1151.2331.1523.22612.6525.26
Wei11.3851.2291.0361.2513.66222.2542.22
Wei31.2981.3010.8991.3344.00730.4455.61
Lgau11.5571.0621.3271.1123.0468.7418.30
Lgau21.5031.1251.2311.1523.24612.6526.04
Lgau31.4341.1911.1201.2083.49218.0735.60
Lgau41.2151.3650.8071.4294.38139.7070.14
Prt1.4631.1591.1641.1843.37715.7331.13

Table 3.

Results for X¯-R chart.

This row gives the generally accepted values of bias correction factors.


Advertisement

6. What happens when the mean shift is transient?

This issue was studied in a paper published in 2021 [42]. We investigated the impact of transient shifts on the operational characteristics of ShCCs. Traditionally, all guides and standards tell the readers about the efficiency of the X-bar chart to detect the shift of the process mean. This efficiency notably surpasses the efficiency of x-chart, and the superiority grows when the sample size n increases. However, this is true but only for the so-called sustained shift. When the shift becomes transient, the situation may change radically, which most practitioners simply do not know because this issue is rarely discussed in the literature. In [42], we discussed an elementary model shown in Figure 3.

Figure 3.

(a) – sustained shift, and (b) – transient shift, n – sample size, here n = 3 The number of subgroups lying entirely inside the step we denoted as m. Here m = 1.

In order to compare the efficiency of different types of ShCCs one needs to calculate the so-called power function (PF) for each type of chart. In Ref. [42] there was considered the case when the transient shift starts between subgroups and ends after m-shifted subgroups within the (m + 1)th. In other words, we assumed that the X-bar chart with subgroups of size n was being built and an assignable cause of variation emerged between subgroups and lasted so that it covered m subgroups totally and r points within the last (m + 1)th. In this case, it was obtained for the average of shifted subgroups:

X¯=[nm(μ0+kσ)+r(μ0+kσ)+(nr)μ0]n(m+1)=μ0+mm+1kσ+rkσn(m+1)E7

Clearly, the bias of the average will be approaching the sustained value when m, and the impact of the transient shift will be maximal for m = 0. In this case the average is equal to

X¯=[r(μ0+kσ)+(nr)μ0]n=μ0+rkσnE8

Obviously, the shift of the average will be decreasing and therefore the probability of detecting it will be decreasing as well. The value of SD is assumed to be constant (at least at the first approximation) and equal to σ/n. Thus, one can obtain:

PF=1Φ(3krn)E9

Using of traditional assumptions of data normality and of being i.i.d. (identically independently distributed) was implied. The results of calculations on the base of (9) for different values of n and r are shown in Figure 3 of [42]. One example of those pictures is given in Figure 4.

Figure 4.

An example of PF for sample of five with different values of r.

As one could expect, when r is equal to n the PF coincides with the traditional one given in many books (see, for example, Figure 10.1 in [9]).

As soon as not all the subgroup’s points fall into the changed process, the PF starts to decrease, and for some value of r – in [42], we called this value boundary, rb – the PF becomes less than the corresponding PF for n = 1 (the PF of the chart for the individual values). It is quite easy to find out the value of rb for which the PF of X-bar chart becomes less than the PF of the chart for individuals:

rbnE10

This means that the traditional conclusion about the superiority of X-bar chart over an X-chart does not work, at least when the duration of the process jump is shorter than the time to gather a subgroup. It all depends on the amount of points falling into the changed part of the subgroup. This example displays that even the simplest ShCCs are not as easy in practice as they may seem. And even for the values of r > rb, the resulting values of probability to reveal the signal about the shift may be essentially less than what is traditionally mentioned in SPC textbooks. For example, for an X-bar chart with n = 10, the probability of detecting the mean’s shift of one sigma falls down from 0.564 when the shift is sustained to 0.078 when only five points are within the changed process, and five points belong to the unchanged process.

Advertisement

7. ShCCs and 2 M’s: metrology and management

In the final section of this chapter, we would like to briefly discuss two areas of people’s activity where ShCCs must be used most often and where they are not practically being used at all. These fields are metrology and management.

Metrology’s primary concern is variation of measurement, so it seems very logical that all metrology-responsible people have to know the basics of SPC and have to be able to construct and interpret ShCCs. Unfortunately, reality looks quite the opposite [43, 44].

Management’s primary concern is variation between people and its impact on the organization. Again, the knowledge of ShCCs looks like it leaves no alternative for managers. Once more, the dark side of reality has beaten us to it [45]. Dr. Deming wrote the honest truth in the foreword to [23]: “The fact is that some of the greatest contributions from control charts lie in areas that are only partly explored so far, such as applications to supervision, management, and systems of measurements, including the standardization and use of instruments, test panels, and standard samples of chemicals and compounds” ([23], ii). A detailed discussion of these issues would require writing another chapter that would most likely be much longer than this one. So we decided to confine ourselves to these ultrashort comments in order to put these topics to the next discussion.

Advertisement

8. Conclusions

We live in a World full of complexity, human irrationality, variability, and, consequently, unpredictability [45]. One of the main features of this world is the ubiquitous presence of variations. Variations differ from each other, so the question arises if one needs to react to them or leave them unnoticed. The ability to answer this question is not something innate to the human being. People should be taught, especially taking into account that knowledge about variability and how to cope against it emerged only after Shewhart’s works of 1931 and 1939 [23, 24]. The tool invented by him – control chart – is the only way to understand if the process is controlled/stable/predictable or not. Without this knowledge, there are no chances to improve any system. Ironically, ShCCs turned out to be very simple technically and rather difficult to use profitably. The reason – contextual knowledge is vital for a genuine understanding of what the charts are trying to tell us. ShCC is a communication tool between a system where a process is going and the process owner. Its construction cannot be totally algorithmized [46]. In other words, everybody in any organization should be taught the art of constructing and interpreting ShCCs. In fact, the reality looks completely different. We started this chapter with the quote of Hoyer, Ellis about quality professionals who do not know the basics of SPC4, and here we will continue citing their phrase “And why should they? Our review of a very large number of SPC textbooks reveals page after page of “cookbook” discussions of practically everything under the sun—with very little discussion on the foundation of SPC ([14], 63)”. Unfortunately, the basics of SPC is a set of rules and recipes which are essentially more narrow than wants of practice. We tried to widen this set in this chapter. To this end, we discussed some uncertainties in the interpretation of assignable causes of variation, and proposed to differentiate between assignable causes that change the system, and those, that do not influence it. Then, we considered the impact of data order on ShCCs and suggested using the average moving range as an indicator of nonrandomness. The next part of the chapter was devoted to the influence of data asymmetry on the limits of ShCC. Recent results for some asymmetrical DFs were delivered. Further, we discussed a simple case of transient jump of process mean and demonstrated that some traditional and established rules should be essentially changed. At last, at the very end of the chapter, we formulated two important issues that should be discussed carefully in the near future. They relate to the link between ShCCs and 2 M’s: management and metrology.

Advertisement

Conflict of interest

The authors declare no conflict of interest.

References

  1. 1. Schindowski E, Schürz O. Statistische Qualitätskontrolle. Berlin: Veb Verlag Technik; 1974
  2. 2. Murdoch J. Control Charts. London: The Macmillan Press, Ltd; 1979
  3. 3. Grant EL, Leavenworth RS. Statistical Quality Control. 5th ed. NY: McGraw-Hill; 1980
  4. 4. Kume H. Statistical methods for quality improvement. The Association for Overseas Technical Scholarship (AOTS). 1985. p. 304
  5. 5. Duncan AJ. Quality Control and Industrial Statistics. 5th ed. Homewood, IL: Irwin; 1986
  6. 6. Wheeler DJ, Chambers DS. Understanding Statistical Process Control. 2nd ed. Knoxville: SPC Press, Inc.; 1992
  7. 7. Rinne H, Mittag H-J. Statistische Methoden der Qualitätssicherung. Fernuniversität-Gesamthochschule-in-Hagen. Deutschland: Fachbereich Wirtschaftswissenschaft; 1993
  8. 8. Alwan LC. Statistical process analysis. Irwin/McGrow-Hill Series in Operations and Decision Sciences. 2000. p. 752
  9. 9. Wheeler D. Advanced Topics in Statistical Process Control. 2nd ed. Knoxville: SPC Press; 2004
  10. 10. Montgomery DC. Introduction to Statistical Quality Control. 6th ed. Jefferson City: John Wiley & Sons; 2009
  11. 11. Balestracci D. Data Sanity: A Quantum Leap to Unprecedented Results. USA: Medical Group Management Association; 2009
  12. 12. ISO 7870-2:2013. Control Charts - Part 2: Shewhart Control Charts. ISO. 2013
  13. 13. Sheremetyeva S, Shper V. Business and variation: Friendship or misunderstanding. Standards and Quality. 2022;2:92-97. DOI: 10.35400/0038-9692-2022-2-72-21 (in Russian)
  14. 14. Hoyer RW, Ellis WC. A graphical exploration of SPC. Quality Progress. 1996;6(Part 2):57-63
  15. 15. Hoyer RW, Ellis WC. Another look at “a graphical exploration of SPC”. Quality Progress. 1996;11:85-93
  16. 16. Hoyer RW, Ellis WC. A graphical exploration of SPC. Quality Progress. 1996;5(Part 1):65-73
  17. 17. Woodall W. Controversies and contradictions in statistical process control. Journal of Quality Technology. 2000;32:341-350. DOI: 10.1080/00224065.2000.11980013 [Accessed: February 20, 2023]
  18. 18. Steinberg D. Industrial statistics: The challenges and the research. Quality Engineering. 2016;28(1):45-59. DOI: 10.1080/08982112.2015.1100453 Accessed: February 22, 2023
  19. 19. Woodall W. Bridging the gap between theory and practice in basic statistical process monitoring. Quality Engineering. 2017;29(1):2-15. DOI: 10.1080/08982112.2016.1210449 Accessed: February 20, 2023
  20. 20. Woodall W. Preprint. Available from: https://www.researchgate.net/publication/366090712_Recent_Critiques_of_Statistical_Process_Monitoring_Approaches_12722 [Accessed: February 21, 2023]
  21. 21. Box G. Scientific Statistics, Teaching, Learning, and the Computer. CQPI Report, No. 146; 1996
  22. 22. Quesenberry C. Statistical gymnastics. Quality Progress. 1998;9:77-79
  23. 23. Shewhart W. Statistical Methods from the Viewpoint of Quality Control. N.Y.: Dover Publications, Inc.; 1939/1986
  24. 24. Shewhart W. Economic Control of Quality of Manufactured Product. Milwaukee: ASQ Quality Press; 1931/1980
  25. 25. Shewhart WD. Deming, and six sigma. In: W. Edwards Deming 2007 Fall Conference, Manuscript No. 187. Available from: http://www.spcpress.com/pdf/DJW187.pdf [Accessed: June 18, 2023]
  26. 26. Adler Y, Maksimova O, Shper V. Shewhart Control Charts in Russia and Abroad: Brief Review of the State-of-the Art (Statistical Aspects). 2011. Available from: http://ria-stk.ru/upload/image/stq/2011/N8/082011-1.pdf (in Russian). [Accessed: June 18, 2023]
  27. 27. Jensen WA, Jones-Farmer LA, Champ CW, Woodall WH. Effects of parameter estimation on control chart properties: A literature review. Journal of Quality Technology. 2006;38:349-364
  28. 28. Chakraborti S, Human SW, Graham MA. Phase I statistical process control charts: An overview and some results. Quality Engineering. 2009;21:52-62
  29. 29. Analysis MS, editor. Reference material. In: Chrysler Group LLC, Ford Motor Company. 4th ed. USA: General Motors Corporation; 2010
  30. 30. Adler Y, Shper V, Maksimova O. Assignable causes of variation and statistical models: Another approach to an old topic. Quality and Reliability Engineering International. 2011;27(5):623-628. DOI: 10.1002/qre.1207
  31. 31. Shper V, Sheremetyeva S, Smelov V, Hunuzidi E. Shewhart control charts – An irreplaceable tool of explanatory data Analysis with underestimated potential. International Journal for Quality Research. 2023;18(2) (In press)
  32. 32. Shper V, Adler Y. The importance of time order with Shewhart control charts. Quality and Reliability Engineering International. 2017;33(6):1169-1177. DOI: 10.1002/qre.2185
  33. 33. Box G, Narasimhan S. Rethinking statistics for quality control. Quality Engineering. 2010;22(2):60-72
  34. 34. Rigdon SE, Cruthis EN, Champ CW. Design strategies for individuals and moving range control charts. Journal of Quality Technology. 1994;26(4):274-287
  35. 35. Kenett RS, Zacks S. Modern Industrial Statistics: With Applications in R, MINITAB and JMP. London: John Wiley & Sons; 2014. p. 9
  36. 36. Nelson LS. Control charts for individual measurements. Journal of Quality Technology. 1982;14(3):172-173
  37. 37. Deming W. In: Orsini J, editor. The Essential Deming. Leadership Principles from the Father of Quality. NY: McGrow-Hill; 2013
  38. 38. Wheeler DJ. Are you Sure we Don’t Need Normally Distributed Data? 2010. Available from: https://www.qualitydigest.com/inside/six-sigma-column/are-you-sure-we-don-t-need-normally-distributed-data-110110.html [Accessed: August 2, 2021]
  39. 39. Shper V, Sheremetyeva S. The impact of non-normality on the control limits of Shewhart’s charts. Tyazheloe Mashinostroenie. 2022;1-2:16-29
  40. 40. David HA. Order Statistics. John Wiley & Sons, Inc.; 1970
  41. 41. Burr IW. The effect of non-normality on constants of X¯ and R charts. Industrial Quality Control. 1967;563:566-569
  42. 42. Shper V, Gracheva A. Simple Shewhart control charts: Are they really so simple? International Journal of Industrial and Operations Research. 2021;4:010. DOI: 10.35840/2633-8947/6510
  43. 43. Adler Y, Shper V. Statistical thinking and metrology: problems and decisions. In: XX IMEKO World Congress. Busan, Republic of Korea: Metrology for Green Growth; September 9-14, 2012
  44. 44. Adler P, Felichkina, Shper. The role of distribution functions in metrology. Journal of Physics: Conference Series;1420:012036
  45. 45. Shper V. Solving 21st century challenges. Quality Progress. 2022;9:36-41
  46. 46. Adler Y. Problems that cannot be solved algorithmically and artificial intelligence. Economics and Management: Problems and Decisions. 2018;7/77(5):17-24. (in Russian)

Notes

  • Here and everywhere below the figure printed in Italic after comma indicates the page in the reference.
  • It is noteworthy that using an appropriate ShCC to analyze stability of key processes is a mandatory requirement in the automotive standard ISO/TS 16949.
  • The reverse ratio of SD to (AMR/d2) was called a stability index in [36] (d2 is one of the constants used to construct ShCC). We think that this name is incorrect, but this is a topic of another paper.
  • We are sure that engineers, managers, CEOs, teachers, physicians, etc. should be added to quality professionals.

Written By

Vladimir Shper, Elena Khunuzidi, Svetlana Sheremetyeva and Vladimir Smelov

Submitted: 27 June 2023 Reviewed: 24 November 2023 Published: 23 December 2023