CHAPTER 2. Shewhart control charts and its application. for rare events in healthcare

Size: px
Start display at page:

Download "CHAPTER 2. Shewhart control charts and its application. for rare events in healthcare"

Transcription

1 CHAPTER 2 Shewhart control charts and its application for rare events in healthcare The control charts considered in this chapter are those that are described in introductory text books on statistical process control for industrial processes. This chapter provides examples of their use for monitoring medical processes, the development and the use of control charts to monitor adverse events in health care. 2.1 Basics of control charts There is extensive interest in the use of statistical process control (SPC) in healthcare. While SPC is part of an overall philosophy aimed at delivering recurrent development, the implementation of SPC usually requires the production of control charts. Since SPC is relatively new to healthcare and does not routinely feature in medical statistics texts/courses, there is a need to explain some of the basic steps and issues involved in selecting and producing control charts. Here I tried to show how to plot control charts and guidance on the selection and construction of control charts that are relevant to healthcare. The primary objective of using control charts is to distinguish between common (chance) and special (assignable) causes of variation. The former is generic to any (stable) process and its reduction requires action on the constraints of the process, whereas special cause variation requires investigation to find the cause and, where appropriate, action to eliminate it. The control chart is one of an array of quality improvement techniques that can be used to deliver continual improvement. Before constructing control charts, it is essential to have a clear aim and clear action plans on how special cause data points will be investigated, or how a process exhibiting only common cause variation might be improved. Mohammed (2004) make

2 sure individuals involved in the project are aware of the SPC approach to understanding variation. Furthermore, the data collection method should be designed to provide a dataset that adequately reflects the underlying process. Issues such as measurement system analysis, sampling, rational subgrouping and operational definitions are relevant to this and have been discussed by Kelly (1999), Carey (2003), Ryan (2000) and Wheeler (1992). In industrial practice, Woodall (2000) found useful to distinguish between two phases in the application of control charts. In phase I, historical data are used to provide a baseline, assess stability, detect special causes and estimate the parameters that describe the behavior of the process. Phase II consists of ongoing monitoring with data samples taken successively over time and an assumed underlying probability distribution which is appropriate to the process. Shewhart control charts are highly recommended for phase I whereas cumulative sum (CUSUM) and exponentially weighted moving average (EWMA) charts have been shown to detect smaller process shifts in phase II, although it is generally acknowledged that a combination of charts is likely to be advantageous (Woodall, 2000). Control charts include a plot of the data over time with three additional linesthe centre line (usually based on the mean) and an upper and lower control limit, typically set at ±3 standard deviations (SDs) from the mean, respectively. When data points appear, without any unusual patterns, within the control limits the process is said to be exhibiting common cause variation and is therefore considered to be in statistical control. However, control charts can also be used to identify special causes of variation. There are several guidelines that indicate when a signal of special cause variation has occurred on a control chart. The first and foremost is that a data point appears outside the control limits. Since the control limits are usually set at 3 SD from the mean, we can state that for a process which is producing normally distributed data (only really demonstrable in artificial simulations) the probability of a point appearing outside the precisely determined phase II control limits is about 1 in 370. Although such a calculation gives some guidance about the statistical performance of this

3 particular rule, it is important not to believe normality as a precondition on the use of a control chart (Wheeler, 1992; Woodall, 2000 & Nelson, 1999). Several other tests can also detect signals of special cause variation based on patterns of data points occurring within the control limits. Although there is disagreement about some of the guidelines, three rules are widely recommended: A run of eight (some prefer seven) or more points on one side of the centre line. Two out of three consecutive points appearing beyond 2 SD on the same side of the centre line (i.e., two-thirds of the way towards the control limits). A run of eight (some prefer seven) or more points all trending up or down. Lee and McGreevey (2002) recommended the first rule and the trend rule with six consecutive points either all increasing or all decreasing. Any test for runs must be used with care. For example, Davis and Woodall (1988) showed that the trend rule does not detect trends in the underlying parameter, and the CUSUM chart can work better than the runs rules in phase II, as shown by Champ and Woodall (1987). Practitioners should note that, as the number of supplementary rules used increases, the number of false alarms will also tend to increase. Another rule is that patterns on the control chart should be read and interpreted with insight and knowledge about the process. This will help the user to identify those unusual patterns that also indicate special cause variation but may not be covered in the three rules above. For example, imagine a process that performs suboptimally every Monday (and in so doing, produces a data point near, but not beyond, the lower control limit). If a day-by-day control chart is plotted, the pattern for Saturday would be repeated every seven data points. Although the three rules above do not capture this scenario, clearly, consideration of the process would raise the question, why Mondays? On the other hand, there is a natural human tendency to see patterns in purely random data. There are many different types of control chart, and the chart to be used is determined largely by the type of data to be plotted. Two important types of data are: continuous (measurement) data and discrete (or count or attribute) data. Continuous data involve measurement: for example, weight, height, blood pressure, length of stay

4 and time from referral to surgery. Discrete data involve counts: for example, number of admissions, number of prescriptions, number of errors and number of patients waiting. For continuous data that are available a point at a time (i.e., not in subgroups) the xmrchart (also known as the individuals chart) is often appropriate. For discrete data, the p- chart, u-chart and the c-chart are relevant. We will see how to construct the xmr-chart, p-chart, u-chart and c-chart using examples. 2.2 Standard Control Charts This section provides example of their use for monitoring quality characteristic of clinical and administrative medical processes given in Morton (2005). This section is concerned with the control chart for monitoring quality characteristic expressed in terms of numerical measurement, count and proportion. (A) Control chart for Numerical data A control chart using sample from a process to monitor its average is called chart and the process variability may be monitored with either S chart, to control the SDs of samples, or an R chart, to control the range of the samples. A description of these charts and instructions for their construction may be found in Montgomory (2005, Chapter 5). Morton (2005) recommended the use of S chart to control process variability. Such charts can be useful for displaying aspects of resource use, such as ward occupancy and pharmacy utilization. In health care, most of the times data having sample size ; that is, sample consist of individual observation. In such situations, the control chart for individual is used. The procedure is illustrated in the following example

5 The chart Suppose that a quality characteristic is normally distributed with mean anad standard deviation, where both and are known. If is a sample of size, then, and where. The xmr-chart consists of two charts: x-chart and mr-chart. The x-chart is a control chart of the n observed values,, and the mr-chart is a control chart of the moving ranges of the data. Consider the data in the Table 2.1, which shows the heart rate readings for a patient in the morning over 30 consecutive days. Since these are continuous (numerical measurement) data, we shall plot them using an xmr-chart. Table 2.1: The Heart rate reading for a patient in the morning over 30 consecutive days Reading number Heart rate (x) Moving range (mr) Reading number Heart rate (x) Moving range (mr) The first steps in producing an xmr-chart first calculate the magnitudes of the differences between successive values of the data (i.e., the moving range). We drop the minus sign as we are only interested in the magnitude of the difference, not the direction. The moving ranges are shown in the Table 2.1. In the second step, to produce the x-chart, plot the heart rate readings (x) on a chart against time. The central line for this chart is given by the mean of x, which is denoted by and it is The control limits for the x-chart are given by the formula:

6 where 12.3 is mean moving range and is empirical constant. The calculated upper control limit and lower control limit are and 61.1 respectively. The x-chart is shown in the figure 2.1(a). In the last step, for producing the moving range (mr) chart, plot the moving ranges against time. The centre line is simply the mean moving range and the upper control limit is given by which gives The lower control limit for a mr-chart is taken to be 0. The mr-chart is shown in the figure 2.1(b). Although the estimation of the process standard deviation (sigma) is based on an underlying normal distribution, it is important to appreciate that the xmr-chart is robust (like other exploratory tools) to departures from the assumption of normality. Thus it is not necessary to ensure that the data behave according to the normal distribution before plotting a control chart. Woodall (2006) noted that, when the data are clearly not normally distributed the statistical performance measures of the charts evaluated under the assumption of normality cannot be trusted to be accurate. One characteristic of time-ordered data that should be assessed is independence over time. Independence means that there is no relationship or autocorrelation between successive data points. With positively autocorrelated data, there will be a tendency for high values to follow high values and low values to follow low values, resulting in cyclic behaviour. With negatively autocorrelated data (less common in applications) high values tend to follow low values and vice versa. Wheeler (1995) shows that even with moderate autocorrelations, the xmr-chart behaves well (i.e., the control limits remain similar), but when the correlation is large (e.g., with a lag 1 correlation coefficient r > 0.6) then the control limits need to be widened by a correction factor. However, Maragah and Woodall (1992) found that even moderate autocorrelation can adversely influence the statistical performance of a chart to the extent that time-series based methods may be more appropriate for autocorrelated data. A common approach is to model the time series and use control charts on the residuals to detect changes in the process

7 Figure 2.1(a): x-chart using the heart rate reading of patients

8 Figure 2.1(b): mr-chart using the heart rate reading of patients. It is possible that sometimes the lower control limit will be below 0. If the data under question come from a process in which negative measurements are not feasible then the lower control limit is routinely reset to 0 to reflect this. It is often stated that data points are needed on an xmr-chart before the limits are sufficiently firmed-up so that if a process is showing common cause variation over these data points then we can confidently conclude that it is stable (Carey, 2003). This is reasonable advice, but it does not imply that a control chart with fewer data points is not useful (Wheeler, 1995). With fewer data points, the resulting control limits are regarded as being soft or provisional control limits. In practice, where there is a stream of data, provisional control limits can be computed and refined (firmed up) with additional data in due course. Another important reason for re-computing the

9 control limits is when knowledge about the underlying process deems it necessary (e.g., a revised definition of the data, material changes to the process, etc.). Jensen et al. (2006) reviewed the literature on the effect of estimation on control chart performance. They showed that sample sizes need to be fairly large to achieve the statistical performance expected under the known parameters case. The xmr-chart is a robust, adaptable chart that has been used with a variety of processes. However, when the underlying data exhibit seasonality or the data are for rare events then preliminary work with the data is required before it can be placed on an xmr-chart (Wheeler, 2003). There is disagreement in the literature on the use of the xmr-chart (Woodall, 2006). Some have argued that the EWMA chart is superior to the xmr-chart in phase II applications. Others have shown that for time-between-events data, or successes between - failures data, other charts suggested by Benneyan (2001) and Quesenberry (2000) are superior. However, the xmr-chart continues to be a useful, if not always optimal, tool for identifying special causes of variation in many practical applications (Wheeler, 1995). Even when the data are from a highly skewed distribution (e.g., exponential waiting times, see Normality section below) a straightforward double square root transformation will often render the data suitable for an xmr- chart (Montgomery, 2005). According to Wheeler (1995, 2003), the xmr-chart is even useful when considering count (attribute) data. It is customary to produce the xmr-chart using the mean (of the data and the moving ranges) statistic for the centre line. However, the median statistic may also be used, but the constants used in the formulas to produce the three-sigma limits will be different (Wheeler, 1992) (3.145 instead of 2.66 for the x-chart and instead of for the mr-chart). Wheeler and Chambers (1992) suggest that when the mean moving range appears to be inflated by just a few data points, xmr-charts based on the median may be a more suitable choice, because the median statistic, although less efficient in its use of the data compared with the mean, affords greater resistance (to extreme data points) than the mean

10 It is sometimes argued that a precondition to the use of a control chart is for assumptions of normality to be met (Nelson, 1999; Wheeler 1995). This is not correct. (This issue has been discussed in depth by Woodall (2000)). Control charts have been shown to be robust to the assumptions of normality. Furthermore during phase I applications the distributional assumption cannot even be checked because the underlying process may not be stable. As special causes of variation are removed the process becomes more stable. Also the form of the hypothesized distribution becomes more apparent and useful in defining the statistical performance characteristics of the control chart. According to Woodall (2006), during phase I applications, practitioners need to be aware that the probability of signals of special cause variation depend primarily on the shape of the underlying distribution, the degree of autocorrelation and the number of samples

11 (B) Control chart for Proportion data In industry, the Shewhart control chart is used to monitor the fraction of items that do not conform to one or a number of standards. Its design and an example showing how it may implemented to monitor the output of an industrial production process is given by Montgomery (2005, Chapter 6). Morton (2005) gives examples of its use to monitor complications that occur in the medical context. It was used to monitor the proportion whose length of stay (LOS) exceeds a certain percentile for the relevant diagnosis related group (DRG), the proportion of surgical patients with surgical site infections, mortality in large general purpose intensive care units in public hospitals, which is about 15%, and the proportion of various hospital record that are completed correctly. Control limits may be set using the normal approximation of binomial distribution but, if the number of trials small, the Camp-Poulson approximation (Gebhardt, 1968), derived from the F-distribution, is a more appropriate approximation. If the expected proportion is less than 10%, Morton et al. (2001) proposed that the Poisson approximation to the Binomial distribution be used for data analysis. The data are grouped into block such that there is one expected complication per block, if the process is in control. Then any special cause variation may be detected by monitoring the complications with Shewhart control charts. This monitoring scheme has been superseded by count data CUSUM charts. Morton (2005) found that control charts such as EWMA or CUSUM charts are better than Shewhart charts for detecting small step changes in any quality measure being monitored. If that measure is a proportion of less than 10% for the in control process, count data CUSUM schemes nay be used to detect step shifts in the proportion. Beiles and Morton (2004, figure 6 and 7) illustrate the CUSUM methods with an example where the rate of surgical site infections are monitored to detect a shift from a surgical site infection rate of approximately 3% for the process in control to a rate of approximately 6% for the process out of control. In the informal CUSUM graph the cumulative numbers of infections are plotted against the number of operations

12 undertaken. An out of control signal occurs if the CUSUM graph crosses above the line given by Infections = 0.06 Operations. In the CUSUM test, the surgical site infection data are grouped into a blocks of 30 operations so that the expected number of infections is approximately one per block. He then implements a CUSUM schemes which is equivalent to the decision interval CUSUM scheme for Poisson data given by Hawkins and Olwell (1998). Ismail et al. (2003) developed a monitoring scheme based on scan statistics which they found sensitive in detecting changes in surgical site infection rates that might not be detected using standard control chart methods such as the CUSUM test. The chart Consider the data in Table 2.2. We are interested in finding whether variation in mortality over time is consistent with common cause variation, we shall use a p-chart. Table 2.2: Number of patients who were admitted with a fractured neck of femur and number who died over 24 consecutive quarters. Quarter Number of patients (n) Number of died (x) Quarter Number of patients (n) Number of died (x) The central line is (overall mean mortality) and the control limits for p-chart are given by the formula:

13 To plot a p-chart for the given data, plot the proportion of deaths (p) on the y axis and the time (quarter) on the x-axis and then add the control limits. This calculation is repeated for each quarter to finally produce the control chart shown in figure 2.2. The control limits thus produced are stepped (see figure 2.2) because they reflect the changes in the sample sizes between quarters. There is no evidence of special cause variation. Improvement in this case requires a closer look at the constraints of the process, perhaps with the aid of a Pareto chart (Ryan, 2000) showing the most frequent reasons for death. Figure 2.2: p-chart based on the proportion of deaths following admission to hospital for patient with a fractured neck of femur. The assumptions of the p-chart are that the events (deaths in our case) are: (i) binary (can only have two states, e.g., alive/dead, infected/not infected, admitted/not admitted, etc.);

14 (ii) have a constant underlying probability of occurring; and (iii) that they are independent of each other. One simple process that can meet all these requirements is the repeated tossing of a coin, but it is indeed impossible for a healthcare process to meet all these assumptions exactly. Nevertheless experience indicates that the p-chart is useful in practical circumstances, even where there is gross departure from the ideal conditions. For instance, Deming (1986) shows an example of inspection data that appeared to have much less variation than expected. The control limits were very wide and the data appeared to hug the central line. This under-dispersion was subsequently traced to an inspector who was too frightened to record the proper failure rates, choosing instead to make up the figures to be just below the minimum target. Clearly, assumptions (ii) and (iii) were being violated. Conversely, when the observed variation is much greater than expected a substantial number of the data points fall outside the control limits. This is termed over-dispersion and often indicates that assumption (ii) and/or (iii) have been violated. While there are statistical techniques for allowing for over dispersion, this technical fix does not itself address the fundamental question of why over-dispersion exists. This requires detective work to understand the underlying process and learn why it is behaving in this way. (Wheeler (1995) recommends using the xmr-chart in this situation.). When using the p-chart to plot percentages, all the proportions, the centre line and the control limits are multiplied by 100. In the case where the upper control limit is above 1, or the lower control limit is below 0, such limits are customarily reset to 1 and 0 respectively, because proportions cannot be negative or larger than 1. We do not have a signal if the observed value was to fall exactly on the reset limit. The standard deviation of a binomial variable is usually derived using, hence giving the control limits as. According to Xie et al. (2002) this approximation is good as long as and as is the case in the worked example above, because under these conditions the binomial distribution is fairly symmetrical and so the Shewhart three-sigma concept works well. However, when these conditions are not satisfied, then a different approach is required. One method is

15 to calculate the exact limits using the probability distribution function of the binomial distribution, which, with modern software such as popular spreadsheet packages and statistical packages, is relatively straightforward. Other approaches involving transformations and regression-based limits have also been documented by Ryan (1997; 2000). A common application of a graph similar to the p-chart is a comparison of the performance of healthcare providers over a fixed period. In this instance there is no time order to the data. The proportions, such as the proportions of readmissions, are plotted as a function of the subgroup sizes (n). The resultant chart is a funnel shape which is visually attractive and statistically intuitive because the funnel shows how the variation due to chance reduces with increasing sample sizes. This avoids the misleading ranking of the providers in league tables and its attendant negative consequences (Adab, 2002; Shewhart, 1980)

16 (C) Control chart for Count data In the industrial context, Shewhart charts are used to monitor number of nonconforming items in sample of constant size n. The number of non-conforming items from an in control process is assumed to follow Poisson distribution with parameter c, if the expected number of non-conforming items is known, or with an estimated parameter, if the number of non-conforming is unknown. The estimated,, may be the observed average of non-conformities in a preliminary data set. If the size of the sample is not constant, the number of non-conforming item may be monitored using Shewhart chart, where the average numbers of non-conformities per inspection unit are monitored. Details of the use and construction of and chart given in Montgomery (2005). The Shewhart charts have been used in the clinical context to monitor monthly counts of bacteraemias where there is no reason to doubt the bacteraemias are independent and it is assumed the counts follow an approximate Poisson approximate (Morton, 2005). Monitoring monthly counts using chart is appropriate, if the monthly number of occupied bed days in the hospital is relatively stable. Where there is a considerable variation in a hospital s occupancy, the appropriate monitoring tool is the U control chart which monitors the number of bacteraemias or infections per 1,000 occupied bed days. The chart Consider the data in Table 2.3, which shows the number of falls in a hospital department over a 13-month period. To investigate the type of variation in these data we shall use a u-chart. To plot a u-chart for the above data, plot the fall rate per patient-day (u) on the y-axis and the time (months) on the x axis. Then add the central line and the control limits as they are calculated. To determine the central line, compute the overall fall rate

17 which is given by the total number of falls divided by the total number of patientdays and it is. Table 2.3: Number of falls in a hospital department over a 12-month period Month Jan Feb Mar Apr May Jun July Aug Sept Oct Nov Dec Number of falls Number of patient days Thus, indicates where to place the central line on the y-axis. To derive the control limits we apply this overall average and the number of days in each time period in turn using the formula: This calculation is repeated for each time period to finally produce the control chart shown in figure 2.3. When the calculated lower control limit falls below 0 it is customarily reset to 0 because count/attribute data cannot fall below 0. As can be seen from figure 2.3, the control limits thus produced are stepped because they reflect the changes in the area of opportunity (number of patient-days) between sampling periods. There is no special cause of variation in months. The assumptions of the u-chart are that the events: (i) occur one at a time with no multiple events occurring simultaneously or in the same location; and (ii) are independent in that the occurrence of an event in one time period or region does not affect the probability of the occurrence of any other event. Although it is rare for all these assumptions to be met exactly in practice, experience indicates that the u-chart is useful in practical circumstances, even where there is a marked departure from the ideal conditions

18 Figure 2.3: u-chart based on number of falls per patient-day over a 12 month period. Typically low frequency events (e.g., number of major complications following surgery) are plotted using u-charts. This is appropriate because the calculations for the control limits are based on an underlying Poisson distribution which is reasonable for infrequent events. To aid communication though, the low frequency event is often re- expressed: for example, using our falls data, in January there was 2 fall in 812 patientas a rate per days, which is falls per patient-day, but this can be re-expressed 100 patient-days by multiplying by 100 (2/812*100=0.246). Such an adjustment will make no difference to the analysis since a u-chart with the y-axis as falls per 100 patient-days will give the same messages as the one with falls per patient-day, though in a unit of analysis which may be more intuitive to healthcare practitioners

19 The chart Consider the data in Table 2.4 which shows the number of emergency admissions over 24 consecutive Mondays (2 January 2006 to 12 June 2006) to one large hospital in India. Table 2.4: The number of emergency admissions over 24 consecutive Mondays. Monday number Number of emergency admissions (c) Monday number Number of emergency admissions (c) If we regard the people in this hospital s catchment area as being the underlying population from which these admissions occur, we can see that (i) the events are relatively low frequency (in comparison with the size of the underlying population) but (ii) the size of that underlying population is unknown. Assuming that the underlying population is large and fairly constant, we can analyse these data using a c chart, which is essentially a u-chart with. For a c-chart, plot the number of admissions (c) on the y-axis and the days (time) on the x-axis, and then add the central line and the control limits as they are calculated. To determine the central line, compute the overall mean number of admissions which is given by the total number of admissions divided by the total number of days and it is. The control limits are set at since, the standard deviation of a Poisson distribution is the square root of the mean. Hence the upper control limits is 98.9 and lower control limit is The final c-chart is shown in figure

20 Figure 2.4: c-chart using the number of emergency admissions on consecutive Mondays. The assumptions of the c-chart are that the events: (i) occur one at a time with no multiple events occurring simultaneously or in the same location; and (ii) are independent in that the occurrence of an event in one time period or region does not affect the probability of the occurrence of any other event. It is impossible for all these assumptions to be met exactly in practice; experience indicates that the c-chartt is useful in practical conditions, even where there is significant departure from the ideal conditions. It is important to note that the u-chart is preferred to the c-chart when sample sizes or areas of opportunity are available and vary considerably. When such an area of opportunity information is not available then

21 the c-chart is the only real option, provided that it is reasonable to assume that the underlying areas of opportunity are relatively large and fairly constant as is the case in our emergency admissions example here. The control limits for a c-chart are derived from. According to Ryan (2000), when as is the case here, the Poisson distribution is roughly symmetrical and so the Shewhart three-sigma concept works well. However, when is not satisfied, then a different approach is required. One method is to calculate the exact limits using the probability distribution function of the Poisson distribution with the help of computer software or statistical tables. Other approaches, using transformations as well as regression-based limits, have also been suggested by Ryan (1997). When the computation of the lower control limit produces a value below 0, which is customarily reset to 0 because the process is not capable of producing a negative count. However, the negative lower control limit is also an indication that the Poisson distribution is not symmetrical and the computation of exact limits is then recommended. Furthermore, according to Wheeler (1995) the c-chart will break down when the mean number of events falls below 1, at which point plotting the time between events can be helpful (Wheeler, 2003). In this worked example we plotted the number of emergency admissions on consecutive Mondays instead of consecutive days. The reason for this is related to rational subgrouping. Some processes are made up of different underlying cause effect subprocesses and an emergency admission to hospitals is one example. So for instance we do not expect the probability of an emergency admission to hospital to be the same for weekends (Saturday/Sunday) and weekdays, nor do we expect Fridays to be the same as other weekdays. In other words, the process exhibits seasonality. In order not to violate the requirements of rational subgrouping, we split the emergency admissions data by day of week and used Mondays as an illustrative example. If a control chart for emergency admissions for consecutive days is required then methods which first deseasonalise the data must be used (Wheeler, 2003).Developing research and practice

22 2.3 Hospital epidemiology and infection control Epidemiology in the broadest perspective is concerned with the study, identification and prevention of adverse health care events, disease transmission and contagious outbreaks with particular focus within hospital on nosocomial infections and infection control. Nosocomial infections essential are any infections that are acquired or spread as a direct result of a patient s hospital stay (rather than being preexistent as an admitting condition), with a few examples including surgical site infection, catheter infection, pneumonia, bacteremia, urinary tract infections, cutaneous wound infections, bloodstream infections, gastrointestinal infections and others. With estimates of the national costs of nosocomial infections ranging from approximately 8.7 million additional hospital days and 20,000 deaths per year (Finison et al., 1993) to 2 million infections and 80,000 deaths per year (Lasalandra, 1998), it is clear that these problems represents quite considerable health and cost concerns. It is also seen that, the number of U.S. hospital patients injured due to medical error and adverse events has been estimated between 7,70,000 and 2 million per year, with the national cost of adverse drug events estimated at $4.2 billion annually and an estimated 1,80,000 deaths caused partly by iatrogenic injury national wide per year (Bates, 1997; Bedell et al., 1991; Bonger, 1994; Brennan et al., 1991; Cullen et al., 1997 and Leape, 1994). The costs of a signal nosocomial infection or adverse events have been estimated both to average between $2,000 and $3,000 per episode. The National Academy of Science, Institute of Medicine recently estimated that more Americans die each year from medical mistakes than from traffic accidents, breast cancer, or AIDS, with $8.8 billion spent annually as a result of medical mistakes (Kohn et al., 1999). Dr. Deming (1992) wrote about the need to view health care as a collection of processes, and the application of quality control charts to hospital epidemiology and infection control had been suggested by Birnbaum (1984). Moreover, various regulatory and accreditation bodies are expressing a growing interest in the concept of variation reduction and the use of control charts. For example: Joint Commission on Accreditation of Health care organizations (JCAHO), the National Committee for

23 Quality Assurance (NCQA), the U.S. center for Disease Control (CDC), the Health Care Financing Administration (HCFA) and others require or strongly encourage that hospitals and HMO s to apply both classical epidemiology and more modern continuous quality improvement (CQI) methodologies to these significant process concerns, including the use of statistical methods such as statistical process control (SPC). For example, the Joint Commission on Accreditation of Health care Organizations (1997) recently stated their position on the use of SPC as follows: An understanding of statistical quality control, including SPC and variation is essential for an effective assessment process Statistical tools such as run charts, control charts, and histograms are especially helpful in comparing performance with historical patterns and assessing variation and stability. Similarly, a recent paper by several epidemiology from the U.S. Center for Disease Control (Martone, 1991) stated that Many of the leading approaches to directing quality improvement in hospitals are based on the principles of W.E. Deming. These principles include use of statistical measures designed to determine whether improvement in quality has been achieved. These measures should include nosocomial infection rates. Conventional epidemiology methods include both various statistical and graphical tools for retrospective analysis, such as described by Mausner and Kramer (1985) and Gordis (1996), and several more surveillance-oriented methods, such as reviewed by Larson (1980). It is worth noting that collectively these methods tend to be concerned with both epidemic (outbreaks) and endemic (systemic) events, with in SPC terminology equates to unnatural (special cause) and natural (common cause) variability, respectively. Several epidemiologists have proposed monitoring certain infection and adverse events more dynamically over time, rather than time statistically, in manners that are quite similar in nature and philosophy to SPC. It is

24 also is interesting that, Dr. Deming (1942) advocated the important potential of SPC in disease surveillance and to rare events. The application of standard SPC methods to healthcare processes, infection control, and hospital epidemiology has been discussed by several authors, including a comprehensive review in a recent series in Infection control and hospital epidemiology by Benneyan (1998). Examples application include medication errors, patient falls and slips, central line infections, surgical complications, and others adverse events. In some cases, however, none of the most common types of control charts will be appropriate, for example due to the manner in which data are collected, pre-established measuring and reporting metrics, or low occurrence rates and infrequent data. One example of particular note is the use of events-between, number-between, days-between, or timebetween type of data that occasionally are used by convention in some healthcare and other settings. Several important clinical applications of such measures are described below, we tried to illustrate appropriate control charts for these cases, such as for the number of procedures, opportunities, or days between infrequent adverse events. SPC charts are chronological displays of process data used to help statistically understand, control and improve a system, here an infection control or adverse event process. Observed process data, such as the monthly rate of infection or the number of procedures between infections, are plotted on the chart and interpreted in a little while after they become available. Three horizontal control lines plotted, which are calculated statistically and help define the central tendency and the range of natural variation of the plotted values, assuming that the rate of occurrence does not change. By observing and interpreting the behavior of these process data, plotted over time on an appropriate control chart, a determination can be made about the stability of the process according to the following criteria. Values that are fall outside the control limits exceed their probabilistic range and therefore are strong indications that non-systemic causes almost certainly exits that should be identified and removed in order to achieve a single, stable, and predictable process. There also should be no evidence of non-random behavior between the limits, such as trends, cycles, and shifts above or below the center line. To help in the

25 objective interpretation of such data patterns, various within limit rules have been defined. See Duncan (1986) and Grant and Leaven worth (1988) for further information about statistical process control in general and Benneyan (1998, 1998 & 1995) for discussion and application of SPC to healthcare processes. Several approaches to applying SPC to hospital infections or others adverse events are possible, dependent on the situation and ranging in complexity and data required. For example, two standard approaches are to use u or p Shewhart control charts, such as, for the number or the fraction of patients, respectively, per time period who acquire a particular type of infection. For example, figure 2.5 illustrates a recent u chart for the quarterly number of infections per 100 patient days, although it should be noted that this example ideally should contain many more subgroups (time period) of data. In terms of proper control chart selection, note that each type of chart is based on a particular underlying probability model and is appropriate in different types of settings. In this particular example, a Poisson based u chart would be appropriate if each patient day is considered as an area of opportunity in which one or more infections theoretically could occur. Alternatively, a binomial based p chart would be more appropriate if the data were recorded as the number of patient days with one or more infections (i.e., Bernoulli trials). Further information on different types of control charts and scenarios in which each is appropriate can be found in the references (Duncan, 1986; Benneyan, 1998; Grant and Leaven worth, 1988). In either of the above cases, note that some knowledge of the denominator is required to construct the corresponding chart, such as the number of patient days, discharges, or surgeries per time period. A more troubling problem is that use of these charts in some cases can result in an inadequate number of points to plot or in data becoming available too infrequently to be able to make rational decisions in a timely manner. For example, recall that a minimum of at least subgroups are recommended to confidently determine whether a process is in statistical control and that, in the case of p and np charts, a conventional rule of thumb for forming binomial subgroups is to pick a subgroup size, n, at least large enough so that np 5. For

26 Figure 2.5: u chart with number of infections per quarter processes with low infection rates, say p < 0.01, the consequence of the above comments can be a significant increase in total sample size across all subgroups and in the average run length until a process change is detected. For example, even for p = 0.01, this translates to 5000 data per subgroup 25 subgroups = 12,500 total data. Furthermore, this can be necessitate waiting until the end of linger time periods (e.g., week, month, or perhaps quarter) to calculate, plot, and interpret each subgroup value, perhaps being too late to react to important changes in a critical process and no longer in best alignment with the philosophy of process monitoring in as real-timee as possible. Note that similar subgroup size rules of thumb exist for c and u charts, and lead to the same general dilemma as for np and p charts, in part because all conventional control charts consider the number of infections or adverse events either at the end of some time period or after some pre-determined number of cases

27 To overcome above mentioned difficulties, the number of procedure, events, or days between infections has been proposed due to easiness of use, more timely feedback, near immediate availability of each individual observation, low infection rates, and the simplicity with which non-technical personnel can implement such measures. Being easy to calculate, a control chart can be updated immediately in real time, on the hospital floor, without knowledge the census or other base denominator. In some settings number - between or time - between types of process measures simply may be preferred as a standard or traditional manner in which to report outcomes. Examples of the traditional use of such measures, Nelson et al. (1995) examined the number of open heart surgeries between post-operative sterna wound infections, Pohlen et al. (1996) analyzed the number of infection free coronary artery bypass graft (CABG) procedures between adverse events, Finison et al. (1993) considered the number of days between Clostridium difficile colitis positive stool assays, Jacquez et al. (1996) analyzed the number of time intervals between occurrences of infectious diseases, Nugent et al. (1994) described the use of time between adverse events as a key hospital-wide metric, Pohlen (1996) monitored the time between insulin reactions, and Benneyan (1998) discussed the number of cases between needle sticks. In order to plot any of these types of data on control charts, however, note that none of the standard charts are appropriate. For example, standard c and u control charts are based on discrete Poisson distribution, np and p charts based on discrete binomial distributions, and and charts are based on continuous Gaussian distributions. By description, conversely, the number of cases between infections most closely fits the classic definition of a geometric random variable, as discussed below and illustrated by the histogram in figure 2.6 of surgeries-between infections empirical data. Note that if the infection rate is unchanged (i.e.; in statistics control), then the probability, p, of an infection occurring will be reasonably constant for each case then this satisfies the definition of a geometric random variable

28 As can be seen by comparison with the corresponding geometric distribution (with the infection probability p estimated via the method of maximum likelihood), these data exhibit a geometric decaying shape that is very close to the theoretic model. Appropriate control charts for these data therefore should be based on underlying geometric distribution, as developed and illustrated below, rather than using any of the above traditional charts. Use of inappropriate discrete distributions and control charts can lead to erroneous conclusions about the variability and state of statistical control of the infection rate, a situation which has been described previously (Benneyan, 1994; Jacquez, 1972; Kaminsky et al. 1992) and is shown in some of the examples in section

29 2.4 g and h control charts The use of control chart techniques for time oriented rare event data is problematical. As the number of events being controlled diminishes, standard control chart techniques such as Individual Value / Moving Range, lose their effectiveness due to the small amount of data. Also with rare event data such as surgical complications, central line infections where the area of opportunity is not unique for each event, the condition for control charting Poisson data or binomial data is violated. In such situation g-type control chart becomes useful. Benneyan (2000, 2001) proposed a g- type control chart based on geometric distribution for monitoring adverse events like number of open heart surgeries between post-operative sternal wound infections, number of time intervals between occurrences of infectious diseases, time between insulin reactions, number of cases between needle sticks etc. Benneyan (1991) and Kaminsky et al. (1992) were developed and investigated g and h charts based on underlying geometric and negative binomial distributions. These new charts were developed specifically to be appropriate for the following three general scenarios: for situations dealing with classic geometric random variables, such as for the various number between type of application for more desirable or powerful control (i.e., greater sensitivity) of low frequency process data than by np and p charts; and If a geometrically decaying shape simply appears naturally, such as periodically observed in histograms of empirical data, even though the existence of a classic geometric scenario may not be apparent. The first two scenarios are of particular interest in this section given the above discussion, with a few examples of the third type of application briefly also illustrated. As shown below, in case with low infection rates and with immediate availability of each observation, consideration Bernoulli processes with respect to geometric rather than traditional binomial probability distributions can produce more plotted subgroup data and greater ability to more quickly detect process changes

30 The mathematical development of g and h chart is given here, because some healthcare practitioners, epidemiologist and other readers might not be familiar with these. Recall that if each case (e.g., procedure, catheter, day etc.) is considered as an independent Bernoulli trial with reasonably the same probability of resulting in a failure (e.g., an infection or other adverse event), then the number of Bernoulli trials units the first failure and the number of Bernoulli trials before the first failure are random variables with Type I and Type II geometric distributions respectively. For example: Let X = number of cases between infections, and p = probability of infection occurring. Here X most closely fits the classic geometric distribution. Since when infection rate is in statistical control, then the probability p will be reasonably constant for each case. Probability of the next infection occurring on the case or immediately after the case are well known to be where a is the minimum possible number of event, a = 1 for Type I (number until) and a = 0 for Type II (number - before) geometric data, with the sum of n independent and identically distributed geometric random variables being negative binomial with probability function Since the event is identical to the event the distribution of is negative binomial distribution. This knowledge is useful in evaluating the probabilities of Type I and Type II errors that can be ascribed to a single subgroup. The expected values, variances, and standard deviations of the total number of cases T and the average number of cases before (a = 0) or until (a = 1) the next n infections or adverse events occur then are