Skip to main content
Erschienen in: Trials 1/2016

Open Access 01.12.2016 | Commentary

Stopping guidelines for an effectiveness trial: what should the protocol specify?

Erschienen in: Trials | Ausgabe 1/2016

download
DOWNLOAD
print
DRUCKEN
insite
SUCHEN

Abstract

Background

Despite long-standing problems in decisions to stop clinical trials, stopping guidelines are often vague or unspecified in the trial protocol. Clear, well-conceived guidelines are especially important to assist the data monitoring committees for effectiveness trials.

Main text

To specify better stopping guidelines in the protocol for such trials, the clinical investigators and trial statistician should carefully consider the following kinds of questions:
1.
How should the relative importance of the treatment benefits and hazards be assessed?
 
2.
For decisions to stop a trial for benefit:
(a)
What would be the minimum clinically important difference for the study population?
 
(b)
How should the probability that the benefit exceeds that difference be assessed?
 
(c)
When should the interim analyses include data from other trials?
 
(d)
Would the evidence meet state-of-the-art standards for treatment recommendations and practice guidelines?
 
 
3.
Should less evidence be required to stop the trial for harm than for benefit?
 
4.
When should conventional stopping guidelines for futility be used for comparative effectiveness trials?
 

Conclusion

Both clinical and statistical expertise are required to address such challenging questions for effectiveness trials. Their joint consideration by clinical investigators and statisticians is needed to define better stopping guidelines before starting the trial.
Hinweise

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

JET conceived the original idea and wrote the first draft of the paper. JET, CP, DW, CD, EB, and AD made critical edits to earlier drafts. All authors read and approved the final version.
Abkürzungen
CID
clinically important difference
NICHD
Eunice Kennedy Shriver National Institute of Child Health and Human Development

Background

Despite long-standing problems in decisions to stop clinical trials [16], stopping guidelines are often vague or unspecified in the trial protocol. The protocol commonly does indicate the number of interim analyses to be performed and the method used to ensure that the type I error does not exceed 0.05 with multiple interim analyses of the primary outcome [79]. However, there is usually little effort to clearly indicate when the findings for both the treatment benefits and hazards [6, 1012] would justify stopping the trial. Moreover, this issue has not been explicitly addressed by the International Conference on Harmonization or such agencies as the Food and Drug Administration [13]. Yet clear, well-conceived guidelines are especially important to assist the data monitoring committees for effectiveness trials [14] intended to determine whether therapies should be used in clinical practice.

Main text

Definitive stopping rules that cover all circumstances are not possible. However, based partly on our experience in the NICHD Neonatal Research Network, we believe better stopping guidelines can be developed if the following basic kinds of questions are jointly and carefully considered by the clinical investigators and statisticians and addressed in preparing the study protocol:
1.
How should the relative importance of the treatment benefits and hazards be assessed?
The importance (value) of treatment hazards relative to the benefits [15, 16] must be judged in defining stopping guidelines for either benefit or hazards. Although difficult to make, these judgments are implicit in any stopping guideline (as well as any treatment recommendation or practice guideline) [1720] and are generally based on the views of investigators or clinicians. Ideally these judgments would be based on prior formal assessments of the views of patients (or their surrogates) who have experienced or are most likely to experience these benefits or hazards. This issue is a fertile area for research in increasing meaningful patient involvement in designing clinical trials [21].
 
2.
What evidence should be required to stop the trial for benefit?
(a)
What would be the clinically important difference (CID)—the minimum magnitude of the treatment benefit large enough to offset the treatment harms [16]—for the patient population in the trial?
The CID depends on the rate and relative importance of treatment benefits and hazards. Clearly and explicitly defined stopping guidelines can specify how the observed rates of treatment benefits and harms can be used in assessing whether the evidence of net benefit or harm is strong enough to justify stopping the trial.
 
(b)
How should the probability that the benefit exceeds the CID be assessed?
While interim analyses are usually performed using only standard frequentist analyses, Bayesian analyses are needed to directly estimate not only the probability of any benefit but also the probability that it exceeds benefit of a specific magnitude such as the CID [2225]. They may also be used if it is appropriate to incorporate data from prior trials of the same intervention to estimate the updated (posterior) probability of benefit based on the current trial. [See item (c).] In addition, Bayesian analyses are also likely to be more understandable than frequentist analyses to clinicians [2326]. For these reasons, Bayesian analyses can be recommended for use with conventional frequentist analyses in assessing interim as well as final results [26, 27]. Even so, judgment will still be required in interpreting the analyses and deciding whether the trial should be stopped.
 
(c)
When should the planned interim analyses include data from other relevant trials?
The effect size associated with low p values in interim results is likely to be larger than that obtained if the trial is continued to its end [1, 2831]. At least for trials where early stopping is considered, consideration of data from prior trials and, if appropriate, incorporation into the interim analyses may well be warranted to avoid premature trial termination and erroneous conclusions due to misleading interim data.
 
(d)
Would the evidence of benefit meet state-of-the-art standards for treatment recommendations and practice guidelines?
These standards address many factors beyond statistical significance [19, 20, 32]. At least for comparative effectiveness trials that compare commonly used therapies and are conducted to provide definitive results to guide clinical practice, a strong case can be made to continue the trials if such standards have not been met despite meeting conventional stopping guidelines [1, 2].
 
 
3.
Should less evidence be required to stop the trial for harm than benefit?
The prudent answer to this question is likely to be “Yes” based on the ethical maxim of primum non nocere and the need to protect patient safety [33, 34] and to focus the limited resources for clinical trials on therapies most likely to be beneficial.
Yet it is uncertain exactly how much evidence of harm should be required, or in Bayesian terms, what probability of net harm is high enough to stop a trial. For treatment harms as for benefits, interim findings extreme enough to be statistically significant or to have high Bayesian probabilities are likely to be more extreme than those obtained if more patients are studied [6, 28, 35, 36]. Misleading findings are especially likely if the analyses are repeated at frequent intervals. Large numbers of future patients, including some patients who otherwise would have been included in the trial, may then be harmed if one erroneously concludes that a truly beneficial therapy is harmful. Quantitative methods to better address this dilemma are now being explored [37].
 
4.
When should conventional stopping guidelines for futility be used for comparative effectiveness trials?
 
Conditional power analyses of interim results or Bayesian predictive probabilities [38] help identify when a significant difference between treatment groups is unlikely to be identified if the trial continues to the preplanned sample size. However, such findings would not necessarily indicate that a trial should be stopped. Even in the absence of significant differences, completion of comparative effectiveness trials comparing two widely used therapies may promote greater use of the therapy with the higher likelihood of benefit, particularly if it is also less invasive, hazardous, expensive, or inconvenient. Moreover, important therapeutic questions are rarely answered in a single trial, and early trial termination will reduce the power of later meta-analyses of all relevant trials to identify important treatment effects for all patients, important subgroups, and patients at differing risk [39]. Such analyses are crucial for applying the results of clinical trials to individual patients [15, 16, 39]—a cutting-edge issue important to augmenting the value of clinical trials.

Conclusions

Although clinical investigators may consider stopping guidelines to be the responsibility of statisticians, both clinical and statistical expertise are required to address such challenging questions. Careful consideration by clinical investigators and statisticians will help to specify better stopping guidelines in the protocol for effectiveness trials, including those that may need especially careful consideration for special populations (e.g., pregnant women or children) or disorders (e.g., cancer). These guidelines can then be reviewed by the data monitoring committee, and any areas of disagreement can be discussed and addressed before starting the trial. As the methods of clinical trials evolve over time, greater input from clinicians and patients will also be needed to promote progressively better informed, better justified, more useful, and more broadly acceptable stopping guidelines.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://​creativecommons.​org/​licenses/​by/​4.​0/​), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://​creativecommons.​org/​publicdomain/​zero/​1.​0/​) applies to the data made available in this article, unless otherwise stated.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

JET conceived the original idea and wrote the first draft of the paper. JET, CP, DW, CD, EB, and AD made critical edits to earlier drafts. All authors read and approved the final version.
Literatur
2.
Zurück zum Zitat Grant AM, Altman DG, Babiker AB, Campbell MK, Clemens FJ, Darbyshire JH, et al. Issues in data monitoring and interim analysis of trials. Health Technol Assess. 2005;9:1–238. iii–iv.CrossRef Grant AM, Altman DG, Babiker AB, Campbell MK, Clemens FJ, Darbyshire JH, et al. Issues in data monitoring and interim analysis of trials. Health Technol Assess. 2005;9:1–238. iii–iv.CrossRef
3.
Zurück zum Zitat Bassler D, Montori VM, Briel M, Glasziou P, Guyatt G. Early stopping of randomized clinical trials for overt efficacy is problematic. J Clin Epidemiol. 2008;61:241–6.CrossRefPubMed Bassler D, Montori VM, Briel M, Glasziou P, Guyatt G. Early stopping of randomized clinical trials for overt efficacy is problematic. J Clin Epidemiol. 2008;61:241–6.CrossRefPubMed
4.
Zurück zum Zitat Guyatt GH, Briel M, Glasziou P, Bassler D, Montori VM. Problems of stopping trials early. BMJ. 2012;344:e3863.CrossRefPubMed Guyatt GH, Briel M, Glasziou P, Bassler D, Montori VM. Problems of stopping trials early. BMJ. 2012;344:e3863.CrossRefPubMed
5.
Zurück zum Zitat Zannad F, Gattis Stough W, McMurray JJV, Remme WJ, Pitt B, Borer JS, et al. When to stop a clinical trial early for benefit: lessons learned and future approaches. Circ Heart Fail. 2012;5:294–302.CrossRefPubMed Zannad F, Gattis Stough W, McMurray JJV, Remme WJ, Pitt B, Borer JS, et al. When to stop a clinical trial early for benefit: lessons learned and future approaches. Circ Heart Fail. 2012;5:294–302.CrossRefPubMed
6.
Zurück zum Zitat Pocock SJ. Current controversies in data monitoring for clinical trials. Clin Trials. 2006;3:513–21.CrossRefPubMed Pocock SJ. Current controversies in data monitoring for clinical trials. Clin Trials. 2006;3:513–21.CrossRefPubMed
7.
Zurück zum Zitat Pocock SJ. Interim analyses for randomized clinical trials: the group sequential approach. Biometrics. 1982;38:153–62.CrossRefPubMed Pocock SJ. Interim analyses for randomized clinical trials: the group sequential approach. Biometrics. 1982;38:153–62.CrossRefPubMed
8.
Zurück zum Zitat O’Brien PC, Fleming TR. A multiple testing procedure for clinical trials. Biometrics. 1979;35:549–56.CrossRefPubMed O’Brien PC, Fleming TR. A multiple testing procedure for clinical trials. Biometrics. 1979;35:549–56.CrossRefPubMed
9.
Zurück zum Zitat Geller NL, Pocock SJ. Interim analyses in randomized clinical trials: ramifications and guidelines for practitioners. Biometrics. 1987;43:213–23.CrossRefPubMed Geller NL, Pocock SJ. Interim analyses in randomized clinical trials: ramifications and guidelines for practitioners. Biometrics. 1987;43:213–23.CrossRefPubMed
10.
Zurück zum Zitat DeMets DL, Pocock SJ, Julian DG. The agonising negative trend in monitoring of clinical trials. Lancet. 1999;354:1983–8.CrossRefPubMed DeMets DL, Pocock SJ, Julian DG. The agonising negative trend in monitoring of clinical trials. Lancet. 1999;354:1983–8.CrossRefPubMed
11.
12.
Zurück zum Zitat Ellenberg SS, Fleming TR, DeMets DL. Data monitoring committees in clinical trials: a practical perspective. Chichester: Wiley; 2003. Ellenberg SS, Fleming TR, DeMets DL. Data monitoring committees in clinical trials: a practical perspective. Chichester: Wiley; 2003.
14.
Zurück zum Zitat Fletcher RH, Fletcher SW. Clinical epidemiology: the essentials. 4th ed. Philadelphia: Lippincott Williams & Wilkins; 2005. Fletcher RH, Fletcher SW. Clinical epidemiology: the essentials. 4th ed. Philadelphia: Lippincott Williams & Wilkins; 2005.
15.
Zurück zum Zitat Sinclair JC. Weighing risks and benefits in treating the individual patient. Clin Perinatol. 2003;30:251–68.CrossRefPubMed Sinclair JC. Weighing risks and benefits in treating the individual patient. Clin Perinatol. 2003;30:251–68.CrossRefPubMed
16.
Zurück zum Zitat Marcucci M, Sinclair JC. A generalised model for individualising a treatment recommendation based on group-level evidence from randomised clinical trials. BMJ Open. 2013;3:e003143.CrossRefPubMedPubMedCentral Marcucci M, Sinclair JC. A generalised model for individualising a treatment recommendation based on group-level evidence from randomised clinical trials. BMJ Open. 2013;3:e003143.CrossRefPubMedPubMedCentral
17.
18.
Zurück zum Zitat McAlister FA, Straus SE, Guyatt GH, Haynes RB. Users’ guides to the medical literature: XX. Integrating research evidence with the care of the individual patient. Evidence-Based Medicine Working Group. JAMA. 2000;283:2829–36.CrossRefPubMed McAlister FA, Straus SE, Guyatt GH, Haynes RB. Users’ guides to the medical literature: XX. Integrating research evidence with the care of the individual patient. Evidence-Based Medicine Working Group. JAMA. 2000;283:2829–36.CrossRefPubMed
19.
Zurück zum Zitat Andrews JC, Schünemann HJ, Oxman AD, Pottie K, Meerpohl JJ, Coello PA, et al. GRADE guidelines: 15. Going from evidence to recommendation—determinants of a recommendation’s direction and strength. J Clin Epidemiol. 2013;66:726–35.CrossRefPubMed Andrews JC, Schünemann HJ, Oxman AD, Pottie K, Meerpohl JJ, Coello PA, et al. GRADE guidelines: 15. Going from evidence to recommendation—determinants of a recommendation’s direction and strength. J Clin Epidemiol. 2013;66:726–35.CrossRefPubMed
20.
Zurück zum Zitat Vandvik PO, Brandt L, Alonso-Coello P, Treweek S, Akl EA, Kristiansen A, et al. Creating clinical practice guidelines we can trust, use, and share: a new era is imminent. Chest. 2013;144:381–9.CrossRefPubMed Vandvik PO, Brandt L, Alonso-Coello P, Treweek S, Akl EA, Kristiansen A, et al. Creating clinical practice guidelines we can trust, use, and share: a new era is imminent. Chest. 2013;144:381–9.CrossRefPubMed
22.
Zurück zum Zitat Fayers PM, Ashby D, Parmar MK. Tutorial in biostatistics: Bayesian data monitoring in clinical trials. Stat Med. 1997;16:1413–30.CrossRefPubMed Fayers PM, Ashby D, Parmar MK. Tutorial in biostatistics: Bayesian data monitoring in clinical trials. Stat Med. 1997;16:1413–30.CrossRefPubMed
23.
25.
Zurück zum Zitat Harrell FE, Shih YC. Using full probability models to compute probabilities of actual interest to decision makers. Int J Technol Assess Health Care. 2001;17:17–26.CrossRefPubMed Harrell FE, Shih YC. Using full probability models to compute probabilities of actual interest to decision makers. Int J Technol Assess Health Care. 2001;17:17–26.CrossRefPubMed
26.
Zurück zum Zitat Parmar MK, Griffiths GO, Spiegelhalter DJ, Souhami RL, Altman DG, van der Scheuren E, et al. Monitoring of large randomised clinical trials: a new approach with Bayesian methods. Lancet. 2001;358:375–81.CrossRefPubMed Parmar MK, Griffiths GO, Spiegelhalter DJ, Souhami RL, Altman DG, van der Scheuren E, et al. Monitoring of large randomised clinical trials: a new approach with Bayesian methods. Lancet. 2001;358:375–81.CrossRefPubMed
27.
Zurück zum Zitat Wijeysundera DN, Austin PC, Hux JE, Beattie WS, Laupacis A. Bayesian statistical inference enhances the interpretation of contemporary randomized controlled trials. J Clin Epidemiol. 2009;62:13–21. e5.CrossRefPubMed Wijeysundera DN, Austin PC, Hux JE, Beattie WS, Laupacis A. Bayesian statistical inference enhances the interpretation of contemporary randomized controlled trials. J Clin Epidemiol. 2009;62:13–21. e5.CrossRefPubMed
28.
Zurück zum Zitat Chalmers I, Sinclair JC. Promoting perinatal health: is it time for a change of emphasis in research? Early Hum Dev. 1985;10:171–91.CrossRefPubMed Chalmers I, Sinclair JC. Promoting perinatal health: is it time for a change of emphasis in research? Early Hum Dev. 1985;10:171–91.CrossRefPubMed
29.
30.
Zurück zum Zitat Hughes MD, Pocock SJ. Stopping rules and estimation problems in clinical trials. Stat Med. 1988;7:1231–42.CrossRefPubMed Hughes MD, Pocock SJ. Stopping rules and estimation problems in clinical trials. Stat Med. 1988;7:1231–42.CrossRefPubMed
31.
Zurück zum Zitat Bassler D, Briel M, Montori VM, et al. Stopping randomized trials early for benefit and estimation of treatment effects: systematic review and meta-regression analysis. JAMA. 2010;303:1180–7.CrossRefPubMed Bassler D, Briel M, Montori VM, et al. Stopping randomized trials early for benefit and estimation of treatment effects: systematic review and meta-regression analysis. JAMA. 2010;303:1180–7.CrossRefPubMed
33.
Zurück zum Zitat DeMets DL, Friedman LM, Furberg CD. Data Monitoring in clinical trials. New York: Springer; 2006.CrossRef DeMets DL, Friedman LM, Furberg CD. Data Monitoring in clinical trials. New York: Springer; 2006.CrossRef
34.
Zurück zum Zitat Piantadosi S. Clinical trials: a methodologic perspective. 2nd ed. Hoboken: Wiley; 2005.CrossRef Piantadosi S. Clinical trials: a methodologic perspective. 2nd ed. Hoboken: Wiley; 2005.CrossRef
35.
Zurück zum Zitat Demets DL. Futility approaches to interim monitoring by data monitoring committees. Clin Trials. 2006;3:522–9.CrossRefPubMed Demets DL. Futility approaches to interim monitoring by data monitoring committees. Clin Trials. 2006;3:522–9.CrossRefPubMed
36.
Zurück zum Zitat Spiegelhalter DJ. Bayesian approaches to clinical trials and health care evaluation. Chichester: Wiley; 2004. Spiegelhalter DJ. Bayesian approaches to clinical trials and health care evaluation. Chichester: Wiley; 2004.
38.
Zurück zum Zitat Saville BR, Connor JT, Ayers GD, Alvarez J. The utility of Bayesian predictive probabilities for interim monitoring of clinical trials. Clin Trials. 2014;11:485–93.CrossRefPubMedPubMedCentral Saville BR, Connor JT, Ayers GD, Alvarez J. The utility of Bayesian predictive probabilities for interim monitoring of clinical trials. Clin Trials. 2014;11:485–93.CrossRefPubMedPubMedCentral
39.
Zurück zum Zitat Vickers AJ, Kattan MW, Sargent D. Method for evaluating prediction models that apply the results of randomized trials to individual patients. Trials. 2007;8:14.CrossRefPubMedPubMedCentral Vickers AJ, Kattan MW, Sargent D. Method for evaluating prediction models that apply the results of randomized trials to individual patients. Trials. 2007;8:14.CrossRefPubMedPubMedCentral
Metadaten
Titel
Stopping guidelines for an effectiveness trial: what should the protocol specify?
Publikationsdatum
01.12.2016
Erschienen in
Trials / Ausgabe 1/2016
Elektronische ISSN: 1745-6215
DOI
https://doi.org/10.1186/s13063-016-1367-4

Weitere Artikel der Ausgabe 1/2016

Trials 1/2016 Zur Ausgabe