Open Journal of Statistics, 2013, 3, 41-54 http://dx.doi.org/10.4236/ojs.2013.34A005 Published Online August 2013 (http://www.scirp.org/journal/ojs) On the Use of Local Assessments for Monitoring Centrally Reviewed Endpoints with Missing Data in Clinical Trials* Sean S. Brummel1, Daniel L. Gillen2 1Harvard School of Public Health, Center for Biostatistics in AIDS Research, Boston, USA 2Department of Statistics, University of California, Irvine, USA Email: sbrummel@sdac.harvard.edu Received June 25, 2013; revised July 25, 2013; accepted August 2, 2013 Copyright © 2013 Sean S. Brummel, Daniel L. Gillen. This is an open access article distributed under the Creative Commons Attri- bution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. ABSTRACT Due to ethical and logistical concerns it is common for data monitoring committees to periodically monitor accruing clinical trial data to assess the safety, and possibly efficacy, of a new experimental treatment. When formalized, moni- toring is typically implemented using group sequential methods. In some cases regulatory agencies have required that primary trial analyses should be based solely on the judgment of an independent review committee (IRC). The IRC as- sessments can produce difficulties for trial monitoring given the time lag typically associated with receiving assess- ments from the IRC. This results in a missing data problem wherein a surrogate measure of response may provide use- ful information for interim decisions and future monitoring strategies. In this paper, we present statistical tools that are helpful for monitoring a group sequential clinical trial with missing IRC data. We illustrate the proposed methodology in the case of binary endpoints under various missingness mechanisms including missing completely at random assess- ments and when missingness depends on the IRC’s measurement. Keywords: Group Sequential; Information; Independent Review; Endpoint; Missing Data 1. Introduction When conducting a clinical trial that utilizes a subclinical and/or subjective primary endpoint it may be necessary to verify the local investigator assignment of the outcome variable. Sometimes this verification is mandated by a regulatory agency or it may be preferred by a study sponsor. The advantage to verify the outcome is that it may decrease misclassification of the outcome in studies performed at multiple sites. As a recent example, consid- er a phase II clinical trial to investigate the efficacy of an experimental monoclonal antibody in combination with chemotherapy in patients with relapsed chronic lympho- cytic leukemia (CLL). A common endpoint in trials tar- geting CLL is a binary indicator of complete response (CR) of disease following the completion of the thera- peutic regime. To standardize the assessment of CR in CLL trials, most studies now use the NCI revised guide- lines for determining CR [1], as shown in Figure 1. It is clear that the CR criteria in Figure 1 are subclinical and subjective in nature, requiring radiographic assessment of lymph node size. In this case, the trial’s primary endpoint may be validated by an independent review committee (IRC). A recent a paper by Dodd et al. [2] reports an ad- ditional seven trials that used an IRC to review the can- cer progression measurements of a local investigator: two renal cell carcinoma studies, one colorectal cancer study, and four breast cancer studies. In the setting of the CLL trial described above, it would not be unusual for an independent data monitoring committee (IDMC) to periodically assess the futility, and possible efficacy, of the experimental intervention through formal hypothesis testing. In this case a group sequential framework would be natural for maintaining frequentist error rates after conducting multiple interim analyses of accruing data. A great deal of research has been conducted in the area of group sequential methods and it is well known that the operating characteristics of a group sequential design depend on, among other things, the exact timing of interim analyses. The timing of se- quential analyses is measured by the proportion of statis- tical information obtained at an interim analysis relative to the maximal information that is anticipated at the final analysis of the trial [3]. Thus it is important to reliably *This research was supported in part by grant P30CA062203 from the ational Cancer Institute (D.L.G.). C opyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 42 Criteria for determining complete response (CR) Absence of lymphadenopathy, as confirmed by physical examination and CT scan (i.e., all lymph nodes < 1.5 cm in diameter). No hepatomegaly or splenomegaly, as confirmed by physical examination and CT scan. Absence of B symptoms. Normal CBC as exhibited by ○ Polymorphonuclear leukocytes ≥ 1.5 × 109/L (without granulocyte colony stimulating factor [GCSF] or other colony stimulating factor support) ○ Platelets > 100 × 109//L (untransfused) or Hemoglobin > 11.0 g/dL (untransfused and without erythropoietin or othercolony stimulating factor support) or Lymphocyte count ≤ 4 × 109/L Bone marrow biopsy indicating that bone marrow is normocellular for age with less than 30% of the cells being lymphocytes and lymphoid nodules absent. Figure 1. Required criteria for determining a complete re- sponse (CR) in chronic lymphocytic leukemia (CLL). estimate statistical information at each interim analysis in order to properly implement and potentially re-power a chosen group sequential design [4,5]. However, when an IRC is used to adjudicate a trial endpoint there may be a subset of individuals who do not have verified IRC mea- surements at the time of an interim analysis because the final assessment of their outcome has yet to be re- turned by the IRC. This results in a portion of trial pa- tients whose primary response from the IRC is missing but whose assessment from the local site (which is typi- cally much quicker to obtain) is known. Relying solely upon validated responses at the time of an interim analy- sis can result in misleading estimates of statistical infor- mation (at best) and opens the possibility of biased esti- mates of treatment effect (at worst) [2,6]. While the local investigator measurements only serve as a surrogate for the IRC outcome measurements, use of this information on observations that are missing validated outcomes may be helpful in estimating statistical information for sam- ple-size recalculations (also known as sample size re- estimation) and for timing future analyses. In the current manuscript we consider the use of in- formation from local assessments when monitoring an IRC validated binary endpoint such as that encountered in the CLL trial described above. This setting allows us to assess the proposed utility of local assessments in es- timating statistical information in clinical trials where a mean-variance relationship exists, and serves as a case study for the importance of information estimation when monitoring a clinical trial with group sequential stopping boundaries. In Section 2 we discuss the importance of accurately estimating statistical information when im- plementing group sequential stopping rules. This section concludes with an example to illustrate the impact that missing IRC data can have on the operating characteris- tics of a group sequential design. In Section 3, we pro- pose missing data techniques to aid in estimating statis- tical information and show how these methods can be used for implementing group sequential tests. In Section 4 we present a simulation study to illustrate the utility of the proposed approach and conclude with a discussion of the challenges of monitoring group sequential clinical trials with IRC validated endpoints. 2. The Role of Statistical Information in Implementing Group Sequential Trial Designs Consider the CLL trial where interest lies in estimating the effect of intervention on the probability of CR (a bi- nary endpoint). Further, suppose that the ratio of the odds of CR comparing intervention to control is used to assess efficacy. Let ki denote the response of individual i in treatment arm k (k = 1 for control, for interven- tion) with associated response probabilities given by Y 2k PrpY1 ikk . The odds of CR for group k is then given by 1 kk k Odds pp, , and the log- odds ratio is given by 1, 2k log Odds 2 Odds 1 . Finally, suppose that the null hypothesis to be tested is 0:0H against the one-sided alternative :0 a H . Now consider a group sequential test of the above hy- pothesis. For testing a one-sided alternative, many com- monly used group sequential stopping rules consider continuation sets of the form , jj Cab such that jj ab for 1, ,jJ analyses. These bounda- ries may be interpreted as the critical values for a deci- sion rule. For instance, in the CLL trial a test statistic less than a would correspond to a decision in favor of su- periority of the intervention while a test statistic exceed- ing b would correspond to a decision of futility re- garding the intervention. Particular families of group sequential designs correspond to parameterized boundary functions that relate the stopping boundaries at succes- sive analyses according to the proportion of statistical information accrued. For instance, in the context of the CLL trial, if we calculate a normalized statistic ˆˆ Var j Z j where ˆ is the maximum like- lihood estimate of the log-odds ratio computed at analy- sis j with corresponding variance ˆ Var , the propor- tion of statistical information accrued at analysis j can be calculated as ˆˆ Var Var Jj where ˆ Var is the variance of the maximum likelihood estimate of the log-odds ratio computed at the final analysis of the trial under a presumed maximal sample size. That is, represents the fraction of total statistical information, defined as the inverse of the variance of the final odds ratio estimate, available from all patients at the time of interim analysis j. It then follows that for some specified parametric functions * , the critical values for a de- cision rule at analysiscan be given by j aj af Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 43 and bj bf. For critical values on the normalized Z-statistic scale, popular examples of * include a one-sided version of the Pocock [7] stopping rule that takes aj G , and a one-sided ver- sion of the O’Brien-Fleming [8] stopping rule that takes bj f 12 aj j fG , , where in both cases the value of G is chosen to maintain a pre-specified type I error rate. bj f The choice of a stopping rule is generally based upon the assessment of a wide range of statistical operating characteristics across multiple candidate designs [3]. In addition to type I error, commonly considered frequentist operating characteristics include power, stopping prob- abilities at each analysis, and average sample size. These characteristics depend on the sampling distribution of the test statistic under a given group sequential sampling design. Unlike a fixed sample design where a single hy- pothesis test is performed after the accrual of all trial data, the sampling density of a test statistic in a group sequen- tial framework not only depends upon the total amount of statistical information accrued over the entire trial but also on the timing of interim analyses as measured by the proportion of the trial’s maximal statistical information, , attained at each interim analysis [3]. Because of this, there are usually at least two complicating factors that must be dealt with during the monitoring of a clinical trial. First, the schedule of interim analyses may not fol- low the schedule assumed during the design of the trial. Often, meetings of an IDMC are scheduled according to calendar time, and thus the sample size available for analysis at any given meeting is a random variable. Si- milarly, accrual may be slower or faster than planned, thereby resulting in a different number of interim analy- ses than was originally planned. Because the exact stop- ping boundaries are dependent upon the number and timing of analyses, either of these scenarios will necessi- tate modifications of the stopping rule. Second, the esti- mate for response variability that was assumed at the design phase is typically incorrect. As the trial progresses, more accurate estimates may be obtained using available data at each interim analysis. In this case, if one wishes to maintain the originally specified power of the trial then updates to the maximal sample size may be neces- sary due to changes in variance estimates. Of course, changes in maximal sample size will result in changes to the proportion of information at all previously conducted analyses. Two ways to adjust for deviations in the timing of planned analyses in order to maintain some of the trial’s original operating characteristics include the error spend- ing approach [9] and the constrained boundaries algo- rithm [5]. First and foremost, these methods are primarily used to maintain the size of the trial (type I error). A choice must then be made as to whether the maximal sample size or the power to detect a clinically relevant alternative should be maintained. Briefly, the constrained boundaries algorithm for maintaining the power of a one-sided group sequential hypothesis test is imple- mented as follows: At the design stage, boundary shape functions are specified as and , where aj f bj f denotes the planned proportion of maximal statisti- cal information attained at interim analysis j, 1, ,jJ , with 1 J . At the first analysis 1 is determined, and stopping boundaries 1 and are computed. A schedule of future analyses, 2 1 a b ,, , which may dif- fer from the originally assumed schedule of analyses is then assumed and a stopping rule using the design para- metric family * (constraining the first boundaries to be 1 and 1) is found which has the desired power. This consists of searching for a new maximal sample size that has the correct type I error and power to detect the alternative for the parametric design family for the as- sumed schedule of interim analyses. At later analyses, the exact stopping boundaries used at previously con- ducted interim analyses are used as exact constraints at those analysis times, and the stopping boundaries at the current and all future analyses as well as the new maxi- mal sample size needed to maintain statistical power are re-computed using the parametric family of designs spe- cified at the design stage and an assumed schedule of future analysis times. Reference [5] notes that when a b aj f and b fj are defined on the type I and II error spending scales, this procedure is equivalent to the error spending approach given in reference [10]. As noted above, in cases where power is to be main- tained the current best estimate of the variance of the response variable at each interim analysis is typically used in place of the variance assumed at the design stage. Use of a more accurate estimate of the response variabil- ity, and hence statistical information, at earlier analyses provides more accurate estimates of the maximal sample size, NJ, at earlier analyses. This will in turn lead to less variation in the relative timing of analyses as the trial proceeds and NJ is updated. In the context of the moti- vating CLL trial the variability associated with a single sampling unit's response is dependent upon the unit’s IRC response probability. Specifically, if ki Y denotes the response of individual i in treatment arm k (k = 1 for control, k = 2 for antibody) then Var 1 ki p k Ypk, where k is the response probability for group k. The result is that biased estimates of response probabilities at an interim analysis will lead to biased estimates of the variability associated with the response variable. To see the implication of this, consider the case where the con- strained boundaries algorithm described above is used to maintain statistical power by updating the trial’s maximal sample size using a biased estimate of response variabil- ity and statistical information. At the time of an interim p Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 44 analysis, missing IRC validated outcomes may be more or less likely to be positive when compared to observed IRC outcomes. This may occur because positive out- comes often require an additional radiologic reading for confirmation, thus leading to a lagged reporting time. In this case, using only data on the available IRC outcomes would lead to downward bias in the event rate, and hence bias in the estimate of statistical information. The end result may be a tremendously (under-) overpowered study depending on the magnitude and direction of the bias. 3. Example of the Impact of Missing Data In this section we demonstrate the impact on group de- sign operating characteristics when the timing of imple- mented interim analyses deviates from the originally planned analysis schedule. Using parameters similar to those that we have encountered in a previously con- ducted CLL trial, we consider a level 0.05 test of the null hypothesis 0:H0 0 against a lesser alternative a:H , where denotes the log-odds ratio com- paring intervention to control. We consider a study de- sign with 95% power for detecting a true odds ratio of 0.65 under an assumed event rate of 0.2 in the control arm. We further consider implementing 4 analyses that are equally spaced in information time. That is, the desired analysis schedule at the design phase is specified by . 0.43 0.25 ,0.5,0.75,1 To illustrate the impact of changing the timing of ana- lyses we consider a shift parameter l so that . Under the alternative hypothesis, Figure 2 depicts the maximal sample size and the average sample number (ASN) for the symmetric O'Brien-Fleming and Pocock designs as the timing of analyses shifts away from the originally desired equally spaced setting . Figure 2(a) shows that the mini- mum ASN attained by the O’Brien-Fleming design oc- curs at values of l between −0.1 and 0.1, while the mini- mum ASN for the Pocock design occurs at approxima- tely . In addition, Figure 2(b) shows that the maximal sample size for the O’Brien-Fleming design is fairly robust to the timing of analyses. It is clear that the ASN and maximal sample size for the Pocock design is more sensitive to shifts in the analysis timing when compared to the O’Brien-Fleming design. This is be- cause the Pocock is far less conservative at early analyses when compared to the O’Brien-Fleming design. 0.25 ,0.5 ,0.75 ,1ll l 0l 0.06 l From Figure 2 it is clear that changes in the timing of analyses will affect the operating characteristics of a sta- tistical design. We now consider a single simulated example to dem- onstrate the implementation of the constrained bounda- ries approach for trial monitoring and how the stopping boundaries of a planned design and an implemented de- (a) (b) Figure 2. Effects of shifting information time for the first three of four analyses on information time on ASN and maximal sample size evaluated under the alternative hy- pothesis ψ = −0.43. The x-axis is the l value in Π = {0.25 + l, 0.5 + l, 0.75 + l, 1}. (a) Effect on ASN; (b) Effect on maximal sample size. sign can differ due to the estimation of information at interim analyses when this approach is utilized. For this example, a shift in the total information schedule from analysis to analysis will be due to an underestimation of a success probability for a binary endpoint, resulting from missing data. When monitoring a clinical trial with an IRC adjudicated endpoint missing data is likely due to lagged IRC response data. As such, IRC outcomes would be more frequently missing at early analyses, with com- plete data at the final analysis. In this case higher bias in Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 45 the estimated probabilities would be seen at earlier ana- lyses. For illustration purposes the example assumes that only those who would have been classified as having an event by the IRC will have the possibility of being miss- ing. The result is that the event probabilities will be un- derestimated at each analysis and these estimates will trend upward from analysis to analysis until the final analysis where complete data will be available on all subjects. Specifically we assume that 39%, 16%, and 3% of IRC endpoints are missing at the first, second, and third interim analyses; and no IRC endpoints are missing at the final analysis. This setting reflects a similar sce- nario to trials we have previously monitored. We focus on a symmetric O’Brien-Fleming stopping rule with 4 equally spaced analyses, allowing early stop- ping for efficacy and futility, and 95% power for detect- ing an odds ratio of 0.65. This design specification re- sults in a maximal sample size of 1819 patients. In moni- toring the trial we consider re-powering the study at each interim analysis using the constrained boundaries ap- proach of [5] as described in Section 2. For this example, at the first interim analysis the estimated event rates are and with a sample size of 436. With these observed estimates the study is then re-powered with a new maximal sample size of 2705 in order to maintain 95% power for detecting an odds ratio of 0.65. This results in a smaller proportion of informa- tion at the first analysis than originally planned (25% to 16%). Using this estimate of information along with the current best estimate of variability, the efficacy and futil- ity boundaries at the first interim analysis are recomputed to be 0.26 and 2.47, respectively, under the pre-specified symmetric O’Brien-Fleming parametric stopping rule. The observed odds ratio at the first analysis, , is 0.86 and this value lies within the continuation region of the stopping rule. At the second analysis, with data now available on 1145 subjects, the observed success prob- abilities are and . These prob- abilities are higher than those observed at the first analy- sis, resulting in a reduction in the re-computed maximal sample size needed to maintain 95% power for detecting an odds ratio of 0.65. The newly re-computed maximum sample size is reduced to 2176, and the percentage of information for the first two interim analyses shifts to 20% for the first analysis and 53% for the second. Con- straining on the first decision boundaries (shown in Ta- ble 1), the efficacy and futility boundaries at the second analysis are now computed to be 0.66 and 0.98, respec- tively. The observed odds ratio at this analysis is 0.81, again implying continuation of the trial. As before, the study is re-powered at the third analysis and then contin- ues to the final analysis where the final sample size is ultimately 1945 subjects. The final sample size is larger than what was assumed at the design stage due to the shifts in the timing of analyses that resulted from under- estimation of the response probabilities at early analyses. 1 1 ˆ0.110 p 1 2 ˆ0.096 p 0.146 1 OR 2 1 ˆ p 2 2 ˆ0.122 p Had unbiased estimates of the success probabilities at early analyses been available, the total sample size for the trial presented in Table 1 would have been much closer to that of the original design specification. Figure 3 shows the estimated information growth curve at each analysis for the trial. As can be seen in this plot, at the first analysis, the information growth has substantially changed from the planned portioning of information. The (a) (b) Figure 3. (a) Estimates of information growth at each anal- ysis. Differences are due to changes estimates of event rates and recalculating maximal sample size. (b) Deviations in ASN due to changes in the proportion of maximal infor- ation as a function of the log-odds. m Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN Copyright © 2013 SciRes. OJS 46 Table 1. Example of planned and implemented stopping boundaries when statistical information is biased due to missing data. The planned design is a one-sided symmetric O’Brien-Fleming design with 95% power for an odds ratio of 0.65. The ob- served design is the implemented design. Π is the (biased) estimated proportion of information. and denote the probability estimates for the control and antibody arms, respectively. ˆ p1ˆ p2 Analysis (j) 1 2 3 4 Planned Design 10.20p, , 20.14p0.65OR Sample Size 454.8 909.61 1364.41 1816.22 Information Fraction j 0.25 0.50 0.75 1.00 Decision Boundary Efficacy (Odds-scale) 0.42 0.65 0.075 0.81 Decision Boundary Futility (Odds-scale) 1.54 1.00 0.86 0.81 Implemented Design Analysis 1 1 ˆ0.110p, , , 2 ˆ0.096p 0.86OR 0.49Z Sample Size 436 1192 1949 2705 Information Fraction j 0.16 0.44 0.72 1.00 Decision Boundary Efficacy (Odds-scale) 0.26 0.61 0.74 0.81 Decision Boundary Futility (Odds-scale) 2.47 1.06 0.88 0.81 Analysis 2 1 ˆ0.146p, , , 2 ˆ0.122p 0.81OR 1.19Z Sample Size 436 1145 1660 2176 Information Fraction j 0.20 0.53 0.76 1.00 Decision Boundary Efficacy (Odds-scale) 0.26 0.66 0.75 0.81 Decision Boundary Futility (Odds-scale) 2.47 0.98 0.86 0.81 Analysis 3 1 ˆ0.165p, , , 2 ˆ0.136p 0.80OR1.63Z Sample Size 436 1145 1631 1945 Information Fraction j 0.22 0.59 0.84 1.00 Decision Boundary Efficacy (Odds-scale) 0.26 0.66 0.77 0.81 Decision Boundary Futility (Odds-scale) 2.47 0.98 0.84 0.81 Analysis 4 1 ˆ0.170p, , , 2 ˆ0.140p 0.79OR1.83Z Sample Size 436 1145 1631 1945 Information Fraction j 0.23 0.59 0.84 1.00 Decision Boundary Efficacy (Odds-scale) 0.26 0.66 0.77 0.81 Decision Boundary Futility (Odds-scale) 2.47 0.98 0.84 0.81 change in information growth is due to a recalculated maximal sample size, but this recalculation was only necessary because of the underestimated probabilities of success. Specifically, the recalculated maximal sample size at analysis one is much larger than the maximal sample size from the original design. This change in the maximal sample size is due to the dependence of the va- riance of the log-odds ratio on the underlying prob- abili-
S. S. BRUMMEL, D. L. GILLEN 47 ties of success. However, at the third analysis, the infor- mation growth is approximately equal to the original design. Ultimately, both the original and observed design have similar maximal sample sizes, but the ASN, as seen in Figure 3 differs substantially. Specifically, the changes in ASN are due to the observed design not following the original intent of having four analyses that are spaced evenly with respect to information time. In turn, the changes in information alter the decision boundaries, as previously discussed. Ultimately, trials with different boundaries and information levels will have different probabilities of stopping at a given analysis, resulting in different operating characteristics. 4. Using Local Investigator Assessment to Monitor Study Data with Missing IRC Assessments Had unbiased estimates of the underlying success prob- abilities been available at early analyses in the previous example the resulting changes to the maximal sample size would have been unnecessary. This would have re- sulted in decision boundaries similar to those originally specified at the design stage. In this section we discuss methods to improve the estimation of information using all of the observed local investigator assessments. When monitoring an IRC-validated primary endpoint, a reasonable approach might perform hypothesis testing using only complete IRC measurements but would use a missing data model that incorporates local investigator assessments in order to estimate response probabilities and hence statistical information. Provided that local assessment is predictive of the IRC-validated outcome, incorporation of local investigator assessment into the estimation of statistical information will result in im- proved estimates of statistical information, potentially minimizing changes to the trial design’s original operat- ing characteristics. Further, by only using the investigator assessment testing is based solely on observed IRC- va- lidated data. For ease of exposition we consider the use of local in- vestigator assessments when an IRC response is missing at a specific interim analysis and drop the analysis sub- script. Assume that at a given interim analysis local in- vestigator assessments are available for nk subjects in group k, and without loss of generality assume that com- plete data are available for the first rk subjects while the remaining kk nr subjects are missing an IRC assess- ment, 1, 2k . For complete pairs let 12 , kiki ki yyy denote the vector of binary local response 1ki and binary IRC response 2ki for subject i in group k, 1,,k ir , 1,k2 . For subjects with only a local as- sessment and no IRC response, let 12 , kiki ki zyz 1, , k , where zki1 is the unobserved IRC response, ir n, 1,k2 . The total data available for group k can then be summarized in the contingency tables provided in Table 2, where for complete cases 001 2 1 11 k r kki i yyn ki 101 2 1 1 k r kki i nyy , , ki ki yki y 011 2 1 1 k r kki i ny , and . 1112 1 k r kki i ny The unobserved cell counts for the incomplete cases are defined analogously as 001 2 1 11 k k n kki ir ki zy , 1012 1 1 k k n kki ir ki zy ki , 011 2 1 1 k k n kki ir zy , and 111 2 1 k k n kkiki ir zy . In the context of the current problem, the common success probabilities 12 12 Pr and Prand, ,0,1, kab kiki ki ki pYaYb ZaYbab must be estimated for study monitoring. Of course we do not observe kab. However, since the local assessments are observed, the marginal totals and M 0k m1k m are Table 2. Aggregated complete and incomplete observations for group k, k = 1, 2 from a clinical trial with lagged IRC response data. Review Type Local Data Complete Cases No Event Event Total IRC Data No Event 00k n 01k n 0k n Data Event 10k n 11k n 1k n Total 0k n 1k n k r Incomplete Cases No Event Event Total IRC Data No Event 00k M 01k M 0k Data Event 10k M 11k M 1k Total 0k 1k kk nr Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 48 known, and conditional on and , kab nkb m ,~ ,MBinppp kabkabkabkbkb m For the remainder of this section we consider three of many possible procedures to estimate when missing IRC data are present at an interim analysis. Once estimated, can then be used to estimate the sampling variability of a response and hence the available statistical information for sample size adjustment and planning of future analy- ses. yz 1 , . ,0,ab 0001 1011 ,,, kkkkk ppppp ˆk p 4.1. Expectation Maximization Algorithm (EM) The EM algorithm [11] is a well-known approach for finding maximum likelihood estimates in the presence of missing data. Briefly, the EM algorithm augments the observed data likelihood with missing data so that maxi- mum likelihood estimates are easily found. That is, we assume an augmented likelihood , k LYZp . We then compute the expected value (E-step) of the log-aug- mented likelihood with respect to the missing IRC data, conditional on the observed data and the current iteration value for k. In the M-Step, the log-augmented likeli- hood is maximized as if the conditional expectations were observed data. The E- and M-steps are repeated until convergence to get our estimate for . p ˆk pk Symbolically, for an initial estimate for k, , the estimate of is updated using the following algo- rithm, p pl k p k p E-Step: , ,log l k l kk k ZY QELY,Z p ppp M-Step: 1argmax , k ll kk Q p pp k p, where Y and Z denote all observed and unobserved data on local and IRC responses. The algorithm is repeated until a distance metric between and is small, and the final estimate for k is given by 1l k pl k p p1 ˆl kk pp. Appendix 7.1 provides more detailed steps of the EM algorithm to maximize a multinomial likelihood to obtain estimates of when there are missing IRC data. k p 4.2. Multiple Imputation Multiple imputation is another natural approach to ac- count for missing data. To perform multiple imputation in the case of missing IRC assessments we can first model the conditional distribution 1, ki k ZYp and impute the missing data from this distribution D times to obtain D estimates of k. In this manuscript we find the condi- tional distribution by using regression estimates from regressing the IRC data on the local investigator data. p The estimator for is calculated from k p 1 1ˆ Dd k d D k pp , where is the dth imputation estimate of . ˆd k pk p Multiple imputation can be carried out in the multino- mial example above by imputing the missing kij values using a binomial distribution. One possibility is to use logistic regression for the imputation model. In this case we fit a logistic model using the complete data with the IRC data as the outcome and the local investigator data as a predictor. Letting z ˆk and ˆk denote the estimated intercept and slope of the fitted logistic regression model for group k, the missing data can be imputed at the indi- vidual level as 2 2 ˆ ˆ 12 ˆ ˆ e ~. 1e kkki kkki y ki kiy Z YyBernoulli 4.3. Complete Case Analysis The last method that we consider is the complete case analysis. This method is the simplest, as it only analyzes the complete data. While this method represents current practice, it assumes that missingness is missing com- pletely at random (MCAR, [12]) and ignores potentially useful information in local response data. In this case, is simply given by ˆkab pkab k nr. 5. Simulation Study In Section 3 we demonstrated that the operating charac- teristics of a group sequential design depend on the tim- ing of interim analyses and showed how changes from planned information can occur when estimates of re- sponse probabilities are biased. In this section we present a simulation study to illustrate the type one error rate, power, ASN, and 75th percentile of the sample size dis- tribution using the three approaches for incorporating local investigator assessments that were described in Section 4. Following from the previous sections, focus is on test- ing the IRC validated log-odds ratio comparing control to antibody in the context of the CLL trial. Specifically, we consider testing a one sided lower alternative with a type one error rate of 5% and 95% power for the design alter- native of −0.43. The stopping rule is taken to be a sym- metric O’Brien-Fleming design with four equally spaced interim analyses that allow for early stopping in favor of futility or efficacy. The simulations are set so that the true odds ratio is 0.65, comparing antibody to control, regardless of whether outcomes are based on the local investigator or the IRC. However, the control arm event rate was assumed to be 0.20 for the IRC and 0.25 for the Local investigator. The missing data was defined differ- ently to illustrate three missing data mechanisms: MCAR, missing at random (MAR), and not missing at random (NMAR). Under MAR the probability of missing an IRC outcome depends on the assigned event assessment of the local investigator. Under NMAR only positive IRC out- Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 49 comes have the potential to be missing. In the MCAR simulation, at the first analysis, the probability of a missing IRC response was taken to be 17.5%. In the MAR simulations, at the first analysis, the probability that a positive IRC response was missing was taken to be 35% if the local investigator response was positive. Lastly, under NMAR, at the first analysis, the probability that a positive IRC response was missing was taken to be 35%, regardless of the local investigator response. Since interim tests are on accumulating data, the proportion of missing responses decreases with each analysis as all of the patients reach the time for evaluation. For the log-odds ratio, in contrast to the binomial va- riance, the variance of the estimator increases as the suc- cess probability moves away from 0.5. Given that the probability of a response was taken to be less than 0.5 in the simulation study, and because observed IRC response rates are biased downwards under the MAR and NMAR setups, the variance of the odds ratio will decrease as the trial continues. Thus a re-powering of the trial will result in an increase in the maximal sample size. However, be- cause an unbounded maximal sample size is unrealistic in practice (a study sponsor is sure to have logistical and financial constraints), the maximal sample size was con- strained so that it would not be larger than 1.25 times the originally planned maximal sample size (Nmax = 1812, ASNnull = ASNalt = 1172). If this restriction is removed, the observed differences between the missing data mod- els would be more extreme. The simulations reflect a scenario where the investi- gators are not expecting any missing information at the design stage of the trial. Thus, at the first analysis, all of the scenarios analyze the data at 0.25 × Nmax. However, due to missingness less than 25% of the originally planned maximal information is observed at the first analysis. The action is then taken to test the data at the current amount of information then recalculate maximal sample size to maintain power and plan for future analy- ses. Results are based upon 10,000 for each scenario. Table 3 depicts the results from the simulation study. Along the rows we consider the three missing data me- chanisms (MCAR, MAR and NMAR). Under the column Future Timing in Table 3, we consider two ways to se- lect the next interim analysis sample size: oversam- pling in anticipation of missing data (“Predict Info”), and ig- noring the possibility of future missing data (“Info N”). The later scenario is included to illustrate that the pri- mary advantage of incorporating local investigator as- sessments is in the sample size computation at the first analysis time. The three considered monitoring strategies are tabu- lated along the columns of Table 3. As can be seen, all of the approaches exhibit the desired type one error rates. However, when the data are NMAR the simulations Table 3. Simulations under MCAR, MAR, and NMAR showing type one error rates, power, ASN, and the seventy fifth per- centile of the sample distribution for the available case analysis, multiple imputation, and the EM algorithm. Results are based on 10,000 simulated trials under each scenario. Information Estimation Complete Cases Multiple Imputation EM Simulation Parameter Future Timing Reject ASN 75% Sample Reject ASN 75% Sample Reject ASN 75% Sample Predict Info0.045 1102 1314 0.045 1101 1314 0.045 1102 1314 Null Info N 0.050 1075 1208 0.050 1077 1206 0.052 1078 1314 Predict Info0.944 1286 1498 0.945 1285 1496 0.944 1286 1496 MCAR Alt Info N 0.951 1260 1420 0.950 1256 1412 0.951 1261 1420 Predict Info0.047 1084 1260 0.047 1055 1274 0.048 1053 1270 Null Info N 0.046 1057 1196 0.047 1040 1204 0.048 1038 1202 Predict Info0.959 1265 1450 0.952 1212 1436 0.953 1211 1436 MAR Alt Info N 0.956 1240 1424 0.951 1188 1366 0.953 1188 1366 Predict Info0.050 1205 1314 0.050 1124 1278 0.051 1125 1280 Null Info N 0.048 1182 1274 0.048 1099 1220 0.049 1100 1220 Predict Info0.972 1340 1718 0.963 1290 1534 0.964 1292 1616 NMAR Alt Info N 0.972 1321 1640 0.964 1271 1638 0.963 1274 1638 Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 50 show that the power is higher than the specified 95%, ranging from 96% to 97%. Next we discuss the efficiency of the EM algorithm and multiple imputation approaches relative to available case analysis. Under the MCAR setting, the sample size statistics are roughly equal across each of the strategies for estimating statistical information (Null: ASNComplete = 1102, ASNMI = 1101, AS N EM = 1102). This is to be ex- pected since the estimates of variability are valid under MCAR for all of the missing data models. In the MAR simulations, the sample size statistics show a larger sav- ings in ASN when local investigator assessments are used to estimate statistical information (Alt: ASNComplete = 1265, ASNMI = 1212, ASNEM = 1211). These differences are due to the fact that the available case statistic tends to overestimate the variability associated with the final test statistic and project future analyses much too far into the future. Similar patterns are observed for the NMAR sce- narios. We note that the lower sample size estimates rela- tive to the “Predict Info” scenario is due to an overall shift in the originally proposed analysis times. 6. Discussion It is becoming increasingly common for regulatory agen- cies to demand independent verification of study re- sponse in clinical trials that utilize a subclinical and/or subjective primary endpoint. Attaining IRC validation in these cases can result in significantly lagged data. The result is that during the monitoring of a trial, IRC-vali- dated data may only be available on a subset of patients which local investigator assessment of the primary out- come is known at the time of an interim analysis. A fur- ther complicating issue is that the observed IRC lag time may be dependent upon the response. For example, posi- tive responses for disease progression in cancer studies may require an additional radiologic reading. This sce- nario can result in biased estimates of the overall re- sponse probability at the time of interim analysis, re- sulting in erroneous changes to the study’s maximal sample size if the study is to be repowered. In the current manuscript, we illustrated issues with the use of local investigator assessments to re-estimate maximal sample size at the time of an interim analysis. Specifically, we considered three different methods for dealing with missing data that can arise when an IRC is used to vali- date local investigator response measurements. We have shown that using local investigator assignment of an outcome variable can be helpful when monitoring a group sequential trial by obtaining more precise esti- mates of information. When testing is based upon only complete cases and local assessments are used to im- prove information estimates, the proposed methods do not affect type one error rates, ASN, or power when missing IRC-validated outcomes are MCAR. However, when missing data are MAR or NMAR, use of local in- vestigator assessments to estimate study response rates for the purposes of recomputing maximal sample size can be helpful in maintaining the planned operating characteristics of the design. In addition, since the true information will be known at the final analysis, type one error rates will be robust when using a miss-specified missing data model. Relative to the complete case analysis, use of local as- sessments for recomputing maximal sample size resulted in generally lower sample sizes (summarized by ASN and the 75th percentile of the sample size distribution) with little observed change in type I and II error rates. This is a result of lower observed event rates due to the missingness mechanism that was considered. In this case, early analyses that only use complete cases would tend to compute large sample size re-estimates to maintain study power while accounting for the low event rate. This, in turn, pushes future analyses back in information time resulting in generally higher sample sizes. In our experi- ence this is a realistic scenario because missing IRC- validated outcomes tend to have a higher probability of being a positive response since these cases generally re- quire more time and additional radiologic readings. The methods presented in this manuscript are easily implemented using any group sequential package that implements the constrained boundaries approach of [5]. One example is the RCTdesign package for the R statis- tical programming language or S+SeqTrial. Example code for computing decision boundaries at the first anal- ysis while updating information using multiple im- puta- tion is presented in the Appendix. The RCTdesign package is freely available by request from the authors of http://www.rctdesign.org. We have only advocated using local assessments to predict study response probabilities in order to obtain more precise estimates of statistical information. Another potential strategy when monitoring a test statistic with missing data is to test the imputed statistic; however, such an approach would be controversial for primary hypothesis testing since final inference would then be dependent upon a correctly specified missing data model. Further investigation of the use of local investigator as- sessments for estimating treatment effect remains area of open research. In addition, priors for the discordance between the local investigator and IRC measurements could be used at the design stage if available to help cor- rect for the issues discussed in this text. This also re- mains an area of open research. REFERENCES [1] B. D. Cheston, et al., “National Cancer Institute-Spon- sored Working Group Guidelines for Chronic Lympho- Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 51 cytic Leuemia: Revised Guidelines for Diagnosis and Treatment,” Blood, Vol. 87, No. 12, 1996, pp. 4990-4997. [2] L. E. Dodd, et al., “Blinded Independent Central Review of Progression-Free Survival in Phase III Clinical Trials: Important Design Element or Unnecessary Expense?” Journal of Clinical Oncology, Vol. 26, No. 22, 2008, pp. 3791-3796. doi:10.1200/JCO.2008.16.1711 [3] S. S. Emerson, J. M. Kittelson and D. L. Gillen, “Fre- quentist Evaluation of Group Sequential Designs,” Statis- tics in Medicine, Vol. 26, No. 28, 2007, pp. 5047-5080. doi:10.1002/sim.2901 [4] D. L. Demets and G. K. K. Lan, “An Overview of Se- quential Methods and Their Application in Clinical Tri- als,” Communications in Statistics-Theory and Methods, Vol. 13, No. 19, 1984, pp. 2315-2338. doi:10.1080/03610928408828829 [5] B. E. Burington and S. S. Emerson, “Flexible Implemen- tations of Group Sequential Stopping Rules Using Con- strained Boundaries,” Biometrics, Vol. 59, No. 4, 2003, pp. 770- 777. doi:/10.1111/j.0006-341X.2003.00090.x [6] R. Ford, et al., “Lessons Learned from Independent Cen- tral Review,” European Journal of Cancer, Vol. 45, No. 2, 2009, pp. 268-274. doi:/10.1016/j.ejca.2008.10.031 [7] S. J. Pocock, “Group Sequential Methods in the Design and Analysis of Clinical Trials,” Biometrika, Vol. 64, No. 2, 1997, pp. 191-199. doi:10.1093/biomet/64.2.191 [8] P. C. O’Brien and T. R. Fleming, “A multiple Testing Procedure for Clinical Trials,” Biometrics, Vol. 35, No. 3, 1979, pp. 549-556. doi:10.2307/2530245 [9] G. K. K. Lan and D. L. DeMets, “Discrete Sequential Boundaries for Clinical Trials,” Biometrika, Vol. 70, No. 3, 1983, pp. 659-663. doi:10.2307/2336502 [10] S. Pampallona, A. A. Tsiatis and K. Kim, “Spending Functions for Type I and Type II Error Probabilities,” Technical Report, Harvard School of Public Health, Dep- tartment of Biostatistics, Boston, 1995 (unpublished). [11] A. P. Dempster, N. M. Laird and D. B. Rubin, “Maxi- mum Likelihood from Incomplete Data via the EM Algo- rithm,” Journal of the Royal Statistical Society, Vol. 39, No. 1, 1977, pp. 1-38. [12] D. B. Rubin, “Inference and Missing Data,” Biometrica Trust, Vol. 63, No. 3, 1976, pp. 581-592. doi:/10.1093/biomet/63.3.581. Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 52 Appendix A1. Steps for the EM Algorithm In the context of using local assessments to estimate IRC response probabilities it is straightforward to compute the conditional expectation of Z given Y. Using the notation of Section 4 and omitting the group indicator, the augmented likelihood is given by, 000001 0110 1011 11 0001 1011 ,nMnM nM nM LYZ pppp p, And log-augmented likelihood is then 00000001010110 101011 1111 log,loglogloglog .LYZn MpnMpnMpnMp p The log-augmented likelihood is linear with respect to ab , ,0,ab 1 , so the expected value is straightforward to compute. Thus, the conditional expectation of the log-likelihood (E-Step) results in 00 00000011010110010101111111 ,loglogloglog , l Qnmppnmppn mppnmpp pp with abab b ppp . For the M-step, maximizing yields, ,l Qpp 1,*,, 0,1. ll ll ababb ababjabb pnmpnnmppnab A2. R 2.14 Code Example ### Load required libraries library(cat); library(RCTdesign) ### Set seed for reproducibility set.seed(1000) ### Helper functions ### Function to calculate odds Odds <- function(x){ x/(1-x)} ### Function to obtain imputation summary impSummary <- function(s,theta){ table(imp.cat(s,theta)[,1])[2]/sum(s$nmobs) } ### Function to perform MI using library(cat) functions GetIRCmi <- function(D,Miss.Data,seed=1){ print('Note: The First Vector in Miss.Data is the IRC Data') ### Pre calculations rngseed(seed) s <- prelim.cat(x=Miss.Data) # preliminary manipulations P <- table(Miss.Data[,1],Miss.Data[,2])/sum(s$nmobs) ### Perform D Imputations p.IRC.imp <- mean(replicate(D,impSummary(s=s,theta=P))) return(list(p.IRC.imp=p.IRC.imp,s=s)) } ### Set parameters for simulated example pCont <- c(IRC=.2,LOC=.25) pTrt <- c(IRC=0.1397849,LOC=0.1780822) OR <- c(IRC=Odds(pTrt)[1]/Odds(pCont)[1],LOC=Odds(pTrt)[2]/Odds(pCont)[2]) ### Use seqDesign() to find initial design Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN 53 dsn <- seqDesign( prob.model = "odds", arms = 2, log.transform =T, null.hypothesis=c(0,0), alt.hypothesis=c(log(OR[1]),0), sample.size=c(.25,.5,.75,1), variance=c(solve(pTrt[1]*(1-pTrt[1])),solve(pCont[1]*(1-pCont[1]))), nbr.analyses=4,test.type="less", size=.05, power=.95, P=1, display.scale = "X",early.stopping="both",design.family="X") dsn ### Sample size per group sampleSize <- dsn$par$sample.size n1 <- ceiling(sampleSize[1]/2) ### Analysis One Sample Size MaxN <- ceiling(sampleSize[4]/2) ### Max Sample Size ### Simulate data for treatment arm Trt.Full.Data<-cbind(IRC=rbinom(n1,size=1,p=pTrt[1]),LOC=rbinom(n1,size=1,p=pTrt[2]))+1 IRC.Miss<-ifelse(rbinom(n1,size=1,p=.175),NA,Trt.Full.Data[,1]) Trt.Miss.Data<-cbind(IRC=IRC.Miss,LOC=Trt.Full.Data[,2]) ### Simulate data for control arm Cont.Full.Data <- cbind(IRC=rbinom(n1,size=1,p=pCont[1]),LOC=rbinom(n1,size=1,p=pCont[2]))+1 IRC.Miss <- ifelse(rbinom(n1,size=1,p=.175),NA,Cont.Full.Data[,1]) Cont.Miss.Data <- cbind(IRC=IRC.Miss,LOC=Cont.Full.Data[,2]) #### Obtain imputed IRC parameter estimates using multiple imputation Trt.Imp <- GetIRCmi(D=1000,Miss.Data=Trt.Miss.Data,seed=1) Cont.Imp <- GetIRCmi(D=1000,Miss.Data=Cont.Miss.Data,seed=1) Trt.Imp.Prb <- Trt.Imp$p.IRC.imp Cont.Imp.Prb <- Cont.Imp$p.IRC.imp nObs.Trt <- Trt.Imp$s$n-Trt.Imp$s$nmis[1] nObs.Cont <- Cont.Imp$s$n-Cont.Imp$s$nmis[1] ### Compute statistical information using imputed estimates ### Note: InfoA1 is very close to (nObs.Cont+nObs.Trt)/(2*MaxN) InfoA1 <- (1/((Trt.Imp.Prb)*(1-Trt.Imp.Prb)*nObs.Trt)+ 1/((Cont.Imp.Prb)*(1-Cont.Imp.Prb)*nObs.Cont))^-1 MaxInfo <- (1/((Trt.Imp.Prb)*(1-Trt.Imp.Prb)*MaxN)+1/((Cont.Imp.Prb)*(1-Cont.Imp.Prb)*MaxN))^-1 ### Use update() to update dsn and loop over the updated design until the sample size converges for( i in 1:3){ dsn <- update( dsn,variance=c(solve(Trt.Imp.Prb*(1-Trt.Imp.Prb)), solve(Cont.Imp.Prb*(1-Cont.Imp.Prb))), sample.size=c(InfoA1/MaxInfo,.5,.75,1),display.scale='Z' ) sampleSize <- dsn$par$sample.size MaxN <- ceiling(sampleSize[4]/2) MaxInfo <- (1/((Trt.Imp.Prb)*(1-Trt.Imp.Prb)*MaxN)+1/((Cont.Imp.Prb)*(1-Cont.Imp.Prb)*MaxN))^-1 print(sampleSize) } dsn ### Log-odds summary statistics p.Est.Trt <- mean(Trt.Miss.Data[,1]-1,na.rm=TRUE) p.Est.Cont <- mean(Cont.Miss.Data[,1]-1,na.rm=TRUE) LogOdds <- log(Odds(p.Est.Trt)/Odds(p.Est.Cont)) varLogOdds <- (1/((p.Est.Trt)*(1-p.Est.Trt)*nObs.Trt)+1/((p.Est.Cont)*(1-p.Est.Cont)*nObs.Cont)) Copyright © 2013 SciRes. OJS
S. S. BRUMMEL, D. L. GILLEN Copyright © 2013 SciRes. OJS 54 zStat <- LogOdds/sqrt(varLogOdds) zStat ### Since zStat=-1.8259 and it is greater than -3.4895 but ### less than 1.8293, continue collecting data but constrain the decision ### boundaries used in this first interim analysis. Note: Using the ### decision boundaries on the test statistics scale is equivalent to ### forming a z-stat using the SE from the imputation probabilities. ### Additional function to obtain EM estimates that can be used in place of GetIRCmi() ### Function to perform EM using library(cat) functions GetIRCem <- function(Miss.Data){ print('Note: The First Vector in Miss.Data is the IRC Data') s <- prelim.cat(x=Miss.Data) # preliminary manipulations P <- em.cat(s) # EM Alg row.names(P) <- c('IRC.No.Event','IRC.Event') colnames(P) <- c('Loc.No.Event','IRC.Event') p.IRC.em <- rowSums(P)[2] return(list(p.IRC.em=p.IRC.em,s=s)) } ### Example of computed EM estimated probabilities for the above example Trt.EM.Prb<-GetIRCem(Trt.Miss.Data) Cont.EM.Prb<-GetIRCem(Cont.Miss.Data)
|