Applied Mathematics
Vol.05 No.21(2014), Article ID:52225,9 pages
10.4236/am.2014.521319

Determining Sufficient Number of Imputations Using Variance of Imputation Variances: Data from 2012 NAMCS Physician Workflow Mail Survey*

Qiyuan Pan, Rong Wei, Iris Shimizu, Eric Jamoom

National Center for Health Statistics, Centers for Disease Control and Prevention, Hyattsville, MD, USA

Email: qpan@cdc.gov

Copyright © 2014 by authors and Scientific Research Publishing Inc.

This work is licensed under the Creative Commons Attribution International License (CC BY).

http://creativecommons.org/licenses/by/4.0/

Received 6 October 2014; revised 29 October 2014; accepted 12 November 2014

ABSTRACT

How many imputations are sufficient in multiple imputations? The answer given by different researchers varies from as few as 2 - 3 to as many as hundreds. Perhaps no single number of imputations would fit all situations. In this study, η, the minimally sufficient number of imputations, was determined based on the relationship between m, the number of imputations, and ω, the standard error of imputation variances using the 2012 National Ambulatory Medical Care Survey (NAMCS) Physician Workflow mail survey. Five variables of various value ranges, variances, and missing data percentages were tested. For all variables tested, ω decreased as m increased. The m value above which the cost of further increase in m would outweigh the benefit of reducing ω was recognized as the η. This method has a potential to be used by anyone to determine η that fits his or her own data situation.

Keywords:

Multiple Imputation, Sufficient Number of Imputations, Hot-Deck Imputation

1. Introduction

As a way of handling missing data, multiple imputation (MI) has been steadily gaining popularity in the past several decades. When it comes to the question of how many imputations are sufficient in applying MI, different researchers have given different answers. Rubin suggested 2 to 5 [1] [2] . Schafer and Olsen suggested 3 to 5 [3] . Graham et al. suggested 20 or more [4] . Hershberger and Fisher suggested that several hundred imputations are often required [5] . Allison suggested that one may need more imputations than what were generally recommended in the literature [6] . The primary reason for the variation in recommended numbers of imputations is that different researchers may use different parameters or statistics and what is good for one parameter or statistic may not be good enough for a different parameter or statistic. For example, Rubin used the “relative efficiency” as defined by him to make the decision [1] , whereas Graham et al. used the testing power to make the decision [4] . When a data user decides to adopt MI, he or she does not have standard guidelines on how many imputations he or she should use for his or her particular data situation.

Variables that need imputation can be different, and the imputation models and/or methods can vary. The resources that can be allocated to MI application can also vary from survey to survey. It may not be practical to recommend a specific number of imputations that can universally fit all imputation models or data and management situations. Instead of attempting to prove or disprove recommendations found in the published literature, the current research examines an empirically-based method to determine the minimally sufficient number of imputations needed.

Let m be the number of imputations and η be the least number of imputations that is sufficient to meet a mini- mum requirement chosen by a data analyst. Among the various published answers to the “how-many-are-suffi- cient” question, the one given by Rubin (1987) [1] appears to be by far the most influential. In Chapter 4 of his classic book on MI, Rubin established that the relationship between, the large sample variance of a point estimator, Q, from a finite m, and, the variance from an infinite m, is

, (1)

where is the population fraction of missing information. From this relationship, the relative efficiency (RE) measured in the units of standard errors from using MI is

. (2)

Based on this RE, Rubin stated: “If, even two repeated imputations appear to result in accurate levels, and three repeated imputations result in accurate levels even when”.

The question is: Is the RE, as defined by Equation (2), the best and only parameter that can be used as the criterion to determine η? Different η recommendations from different researchers (e.g. [3] -[5] ) suggest this RE may not be the best and only parameter to determine η. None of these researchers including Rubin have explicitly claimed that his or her method is the only correct method to determine η.

The number of imputations affects the MI results in multiple ways as indicated by the following equations for MI data analyses [1] . The mean of the point estimator Q is

. (3)

The average variance estimate of Q over m complete datasets from MI is

, (4)

where is the variance estimate of the ith imputation. The estimated imputation variance is

. (5)

The total variance estimate is

. (6)

Equations (3) to (6) all contain m as a factor, indicating that m can affect the results of MI data analyses in multiple ways. The minimally sufficient number of imputations could be defined as the smallest m that would produce a sufficiently accurate, , B, or T as judged by the data user.

This paper presents a methodology to determine η by focusing on the effects of m on the accuracy of B as defined by Equation (5). The focus is on B because the primary advantage of MI over single imputation (SI) is that MI makes it possible to estimate B while SI cannot [1] [7] . For an m to become the η, this m should first be capable of allowing us to obtain a sufficiently accurate B. The accuracy of B can be measured by ω, the estimated standard error of B. For a particular variable, ω would be smaller as m increases. If ω becomes sufficiently small, then one may conclude that B is sufficiently accurate and the m corresponding to that ω value would represent η. Using the data from the 2012 wave of the National Ambulatory Medical Care Survey (NAMCS) Physician Workflow study (PWS), we were able to demonstrate that ω can be used to determine η. This method can be used by anyone to determine η for his or her own data.

2. Methodology

2.1. The Survey

The PWS is a nationally representative, 3-year panel mail survey of office-based physicians conducted by the National Center for Health Statistics [8] . The PWS sample includes 5266 physicians who were confirmed eligible in the 2011 Electronic Medical Records mail survey, a supplement to the NAMCS. To meet eligibility criteria, physicians had to see ambulatory patients in office-based settings. All eligible physicians were mailed the first wave of the study in 2011. Sampled physicians who did not respond to the first mailing were sent up to two additional mailings, and survey administrators followed up by phone with those who did not respond to any of the three mailings. A total of 3180 eligible physicians responded in the 2011 PWS, yielding a weighted response rate of 46.0 percent. All eligible physicians, including those who did not respond in 2011, were mailed a second survey in 2012. This 2012 cycle of data yielded 2567 eligible responses, for a weighted response rate of 42.1 percent and was used for the current MI research. Missing value percentages were calculated by regarding the 2567 records as the complete data set and the imputations were carried out within these 2567 records. All analyses in this research were conducted with unweighted data.

2.2. The Imputed Variables

When resources permit, it is desirable to include more variables of different characteristics to make the conclu- sions from the study have better reference values to other data users and researchers. The variables selected for imputations in this study are described in Table 1. Five variables were selected from the survey for MI tests. PRACSIZE2, PRACSIZE5, PRACSIZE20, and PRACSIZE100 represent the number of physicians in the practice. PRACTSIZE100 is the original variable with values ranging from 1 to 100. PRACSIZE2 and PRACSIZE5 were derived from PRACTSIZE100 by recoding the values into 2 and 5 categories, respectively, and PRACSIZE20 was derived from PRACTSIZE100 by top-coding the >20 values to 20. These four variables had

Table 1. Characteristics of variables used.

The same percent of missing data of 3.62% but different value ranges and variances. The fifth variable, CLSTAFF1, was the number of clinical staff. It had a value range of 0 - 99, which was similar to that of PRACTSIZE100, but a missing data percentage of 8.88%, which was different from that of the other four variables.

2.3. The Imputations

Hot deck imputation [9] was used in this MI study. The donor groups for the hot deck imputation were defined by the region of the physician’s interview office (REGION), the physician’s interview specialty group (SPECR), and the physician’s primary present employment code (PRIMEMP) (Table 1). These three variables were chosen for defining the donor groups in order to minimize possible correlation between the donor groups and the variables to be imputed. The donors for missing values were randomly selected with replacement from the donor group which matched the recipient. A total of 500 imputations were obtained for each variable. The pool of 500 imputations was used as the population and a random sample of m imputations was drawn from the pool for the MI.

2.4. Determination of the Variance of the Imputation Variances

Thirteen different numbers of imputations were tested for MI: m = 2, 3, 5, 10, 15, 20, 25, 30, 35, 40, 60, 80, and 100. The imputation variance (B), defined by Equation (5), was calculated for the data obtained from each MI. In order to calculate the variance of B, 10 independent random samples from the 500 imputations were pulled for each m of each variable. Figure 1 describes the process in a diagram. VB, the variance of B, is estimated by Equation (7):

, (7)

where n is the number of MI samples for a given m, which is 10 for the current study. The standard deviation of B is:

. (8)

The standard error of B was calculated using the following formula:

, (9)

VB, SDB, and ω measure the variance of imputation variances at different scales.

2.5. Determination of Sufficient Number of Imputations (η)

For a given m-ω curve, different data users may have different criteria for a particular m value to be recognized as being “sufficient” and so arrive at with different η values. Two methods were developed to determine η, one

Figure 1. Experimental design for obtaining B, the imputation variance, and ω, the standard error of B, for different m, the number of imputations.

is called “the moving regression method” and the other is called “the confidence interval method”. These two methods do not exhaust the possibilities of other ways for using the m-ω relationship to determine η.

2.5.1. The Moving Regression Method

The linear model Y = βX + a was used to fit the m-ω relationship in a “moving” fashion, where the independent variable X is m and the dependent variable Y is ω. Let be the values of the independent variable m arranging from the smallest to the largest and be the corresponding values of ω. Let h be the number of independent variable values included in each regression group. Fit the regression model for, then for, then for, and so on. The relative slope for 5 − m increment (RS5) is derived and a cutoff point is chosen for RS5 to determine η. RS5 is defined as:

, (10)

where S5 is the slope of the regression line using an increment of five in m as a unit of X, and is the mean of ω in the particular regression group. The S5 is the standardized ω reduction slope for every 5-imputation increase in m. RS5 is the percentage of S5 over. The worksheet for variable PRACSIZE5 is given in Table 2 to illustrate this method.

Within the range of m where m has a detectable effect on ω, RS5 should be negative, and the absolute value of RS5 should decrease with the increase in m. With only three data points included in a regression, however, a single outlier of ω may result in outlier RS5 values. Suppose is the cutoff point to determine η. An is called an outlier if it satisfies the following two conditions: 1) It is less than the cutoff point; 2) There are values that are greater than the cutoff point within the next three RS5 values upstream, i.e. in the direction of higher m.

Going from the lowest m to the highest m, if there are no outliers, the middle point m of the first regression with would be the η. In the current case, an outlier ω at m = 25 (Table 2) has resulted in two outlier RS5 values, −1.03 (from the regression of m = 15, 20, 25) and −9.03 (from the regression of m = 20, 25, 30). The following option is suggested to deal with outliers: Let m be the m corresponding to the outlier ω. Take the av- erage of the two RS5 values whose corresponding regressions have om as their largest and smallest m, respec- tively. If the average is less than the cutoff point, then the om would be the η. If the average is

Table 2. The worksheet of the moving regression method (linear model: ω = βm + a) for determining the minimum sufficient number of imputations (η) using PRACSIZE5 as an example.

aThe standardized regression slope using an increment of 5 in m as a unit; bThe mean of ω values included in the regression; cThe percentage of S5 divided by.

greater than the cutoff point, then all the values related to the outlier ω are treated as being greater than the cutoff point. In the current case, the two regressions with the om as their largest and the smallest m have m = 15, 20, 25 and m = 25, 30, 35 and the corresponding RS5 values −1.03 and −27.79, respectively. The average is 14.41 > 10. Therefore, neither −1.03 nor −9.03 would be recognized as the first values. The regression of the first has m = 40, 60, 80. Therefore, the η is determined to be 60 for PRACTSIZE5 using this method.

2.5.2. The Confidence Interval Method

In the confidence interval method, ω is used to calculate the 95% confidence interval for, the mean of the 10 B values of each m. Because is the sample mean of B which is a chi-square variable with m ? 1 degrees of freedom, the normal distribution can be used to approximate the distribution of for sufficiently large values of m [10] . For PRACTSIZE5, m = 2 was verified to be sufficiently large for to approach normality (not shown). As a result, for a sample size of 10, the Student t test can be used to calculate the confidence interval of. Then P, the percentage of half-width of the confidence interval divided by is derived and used as the parameter to determine η. P is defined by the following equation:

, (11)

where t0.05 is the t value from the Student t test at 0.05 probability level and the term “t0.05ω” represents half the width of the 95% confidence interval because the full width of the 95% confidence interval would be . The variable PRACSIZE5 was used to illustrate this method and the results are presented in Table 3. Data in Table 3 show that P decreased as m increased. It is up to each data analyst to decide the cut-off point at which the m would be recognized as sufficiently large so that this m would become η. The minimum m value above the chosen cutoff point of P would be the η. For illustration, then η is determined to be 40 for PRACTSIZE5 if P = 15 is chosen as the cutoff point.

3. Results

3.1. Effects of m on B

The magnitude of B, the imputation variance as defined by Equation (5), was different among the five variables

Table 3. The worksheet of the confidence interval method for determination of the sufficient number of imputations (η) using PRACSIZE5 as an example.

a = The mean of B (imputation variance).

(Table 4). For example, at m = 100, the mean of the B was 0.101, 1.02, 15.6, 186, and 232 for variables PRACSIZE2, PRACSIZE5, PRACSIZE20, PRACSIZE100, and CLSTAFF1, respectively (Table 4). The difference in B values reflected the huge difference in the variances among the five variables. Within each variable, the mean of B did not show a clear trend of increase or decrease as m was increased from 2 to 100 (Table 4). This was expected because the sample mean of B would fluctuate around the population B value, which should not be affected by the changes in m.

Individual B values at each m were plotted in Figure 2. Within each variable, the individual B values were scattered across a wide range when m was small (<20) (Figure 2). For example, for variable PRACSIZE5 at m = 3, the B value was 0.2186 from one sample, and 4.0896 from another sample. This means that if one decided to use m = 3 for MI, the imputation variance could by chance be many times bigger or smaller than the true B value. The widely scattered data points at low m values in Figure 2 indicate that when m was less than 20 or even 40, the B value obtained may not be accurate and reliable.

3.2. Effect of m on ω

Figure 3 presents data on ω for different m values. Measured in standard error, ω was the largest when m = 2 or 3. It quickly decreased with the increase in m and tended to stabilize as m approached 100, the highest m tested.

Table 4. The effects of mean of the imputation variances from the 10 samples at different number of imputations (m).

Figure 2. Effects of number of imputations (m) on imputation variances (B).

Taking the variable PRACSIZE5 as an example, ω decreased by 82% when m increased from 2 to 15, then decreased an additional 13% when m increased from 15 to 100. The m-ω curve was similar among all the five variables tested even though these variables were different in their value range and other characteristics (Figure 3; Table 1).

3.3. Sufficient Number of Imputations

Sufficient numbers of imputations (η) as determined by the moving regression method (see Section 2.5.1) and the confidence interval method (see Section 2.5.2) for the five variables tested are listed in Table 5. The two methods often gave different η values for the same variable from the same m-ω relationship. To be specific, the moving regression method resulted in an η of 30 to 80, whereas the confidence interval method resulted in an η of 25 to 40 (Table 5). These results suggest that the decision on criteria for what should be considered “sufficient” plays an important role in determining η.

4. Discussions

Shifting from SI to MI allows an estimate of B, the imputation variance, and a more accurate estimate of the total variance, T, as defined by Equation (6) [1] [7] . If an MI protocol cannot give a reliable estimate of B, then the

Figure 3. Effects of number of imputations (m) on the variance of the imputation variances (ω).

Table 5. Sufficient numbers of imputations (η) as determined by both the moving regression method (Section 2.5.1) and the confidence interval method (see Section 2.5.2).

major benefit of MI would be lost. Therefore, it makes sense to use an m-dependent measure of reliability for the estimate of B as a criterion in determining η, the minimally sufficient number of imputations. The results of this research indicate that B values obtained for m < 20 may not be reliable at least for these five variables tested using the imputation method described (Figure 2 and Figure 3). If one uses Rubin’s recommendation of m = 2 or 3, the B value obtained may be many times bigger or smaller than the true B value. Using the criterion of whether a reliable B can be obtained using the protocol described in this paper, the recommended m = 2 to 5 ([1] [2] ) are not sufficient for the five variables tested.

The variance of imputation variances is a good determiner of η for three reasons. First, it is critical for any MI procedure to produce a reliable estimate of B. Second, the effect of m on ω can be big and can be easily visualized, as shown in Figure 3. Furthermore, the effect of m on ω decreases with increased m, allowing the data user to set a cutoff point for an m to be recognized as the η. Third, the method of calculating ω is relatively simple and practical.

To determine the minimally sufficient number of imputations, one has to first define the criteria for being recognized as “sufficient”. Different data users may have different criteria. As a result, different η values may be obtained from the same m-ω relationship data, as exemplified by the results show in Table 5 for the two different η determination methods discussed above. This is another reason why a universal recommendation of η is not a good idea and may even be impossible.

The highest m value tested in this study was 100. There was an indication that even m = 100 still had some effect in reducing ω (Figure 3(b), Figure 3(c) and Figure 3(d)). Ten samples were pulled for each m. If one pulled 100 samples for each m and tested m values of 200, 300, or even greater, it may be possible to show that an increase in m would still be statistically effective in reducing the variance of B. Theoretically the effects of m in reducing ω would approach zero but may never become zero. Therefore, the argument by some researchers ([5] ) that η should be in hundreds is not totally unreasonable and has some support from this research. But because an increased m means a higher cost, it may not be justified to do many more imputations just for a tiny bit of gain. Balance of the gain against the cost based on one’s own situation plays an important role in η determination.

Based on discussions above, η can be defined as the value of m beyond which the cost of further increase in m would outweigh the benefit of reducing ω. Both the moving regression method and the confidence interval method were used to determine η based on the m-ω relationships obtained for five survey variables. The two methods are meant to be examples to illustrate that different η values may be obtained depending on the chosen method and the cutoff point. There may be other methods to determine η from a m-ω curve that fit one’s particular situation better. With the ever-improving information technology, the cost for increased number of imputations may not be a major source of cost in determining η. But there is always a need to consider one’s particular situation such as one’s tolerance level for the probability in determining η.

The five variables are different in their value ranges, variances and missing data percentages. Nevertheless the m-ω relationship curve can be used to determine η for all of the five variables tested. No obvious relationship can be established between the variable characteristics such as the variances and the η values obtained.

5. Conclusions

5.1. The Variance of Imputation Variances Is a Good Determiner for Data―Specific η

The primary advantage of MI over SI is that MI makes it possible to obtain an estimate of B while SI cannot. Ideally m should be large enough to ensure a reliable estimate of B, otherwise the MI advantage would be compromised. The reliability of B can be measured by ω. In this study, the ω decreases as m increases and the unit gain from increased m decreases with greater m. These characteristics of the m-ω curve make ω an ideal determiner of η. The method described in this paper can be used by any data user to determine the η that fits his or her particular data situation.

5.2. Sufficient Number of Imputations Is Larger Than Popular Recommendations

The most popular recommendation for η is between 2 and 5, suggested by Rubin [1] [2] . Our results indicate that 2 - 5, or even 10, imputations may not be sufficient to obtain a statistically reliable B in MI data analyses at least for the five survey variables tested in this research.

Acknowledgements

The NAMCS Physician Workflow mail survey is sponsored by the Office of the National Coordinator for Health Information Technology.

References

  1. Rubin, D.B. (1987) Multiple Imputation for Nonresponse in Surveys. John Wiley & Sons, New York, 1-23, 75-147.
  2. Rubin, D.B. (1996) Multiple Imputation after 18+ Years (with Discussion). Journal of the American Statistical Asso- ciation, 91, 473-489. http://dx.doi.org/10.1080/01621459.1996.10476908
  3. Schafer, J.L. and Olsen, M.K. (1998) Multiple Imputation for Multivariate Missing Data Problems: A Data Analyst’s Perspective. Multivariate Behavioral Research, 33, 545-571. http://dx.doi.org/10.1207/s15327906mbr3304_5
  4. Graham, J.W., Olchowski, A.E. and Gilreath, T.D. (2007) How Many Imputations Are Really Needed? Some Practical Clarifications of Multiple Imputation Theory. Prevention Science, 8, 206-213. http://dx.doi.org/10.1007/s11121-007-0070-9
  5. Hershberger, S.L. and Fisher, D.G. (2003) A Note on Determining the Number of Imputations for Missing Data. Structural Equation Modeling. Structural Equation Modeling, 10, 648-650. http://dx.doi.org/10.1207/S15328007SEM1004_9
  6. Allison, P. (2012) Why You Probably Need More Imputations Than You Think. http://www.statisticalhorizons.com/more-imputations
  7. Rubin, D.B. (1978) Multiple Imputations in Sample Surveys―A Phenomenological Bayesian Approach to Nonresponse. Proceedings of the Survey Research Methods Section of the American Statistical Association, 20-28.
  8. Jamoom, E., Beatty, P., Bercovitz, A., et al. (2012) Physician Adoption of Electronic Health Record Systems: United States, 2011. National Center for Health Statistics, Hyattsville, NCHS Data Brief, No. 98.
  9. Andridge, R.R. and Little, R.J.A. (2010) A Review of Hot Deck Imputation for Survey Non-Response. International Statistical Review, 78, 40-64. http://dx.doi.org/10.1111/j.1751-5823.2010.00103.x
  10. Freedman, D.R., Pisani, R. and Purves, R. (2007) Statistics. 4th Edition, W. W. Norton & Company, New York, 415-424, 488-495, 523-540.

NOTES

*The findings and conclusions in this paper are those of the authors and do not necessarily represent the views of the National Center for Health Statistics or the Centers for Disease Control and Prevention of the United States government.