Identification of Clinically Useful Cancer Prognostic Factors: What Are We Missing? (original) (raw)

Journal Article

,

Affiliations of authors: Biometric Research Branch, National Cancer Institute, Bethesda, MD (LMM); Cancer Research UK/NHS Centre for Statistics in Medicine, Oxford, UK (DGA); Institute for Medical Biometry and Informatics, University Hospital Freiburg, Freiburg, Germany (WS)

Correspondence to: Lisa McShane, PhD, National Cancer Institute, MSC 7434, 6130 Executive Blvd., Bethesda, MD 20892-7434 (e-mail: lm5h@nih.gov ).

Search for other works by this author on:

,

Affiliations of authors: Biometric Research Branch, National Cancer Institute, Bethesda, MD (LMM); Cancer Research UK/NHS Centre for Statistics in Medicine, Oxford, UK (DGA); Institute for Medical Biometry and Informatics, University Hospital Freiburg, Freiburg, Germany (WS)

Search for other works by this author on:

Affiliations of authors: Biometric Research Branch, National Cancer Institute, Bethesda, MD (LMM); Cancer Research UK/NHS Centre for Statistics in Medicine, Oxford, UK (DGA); Institute for Medical Biometry and Informatics, University Hospital Freiburg, Freiburg, Germany (WS)

Search for other works by this author on:

Cite

Lisa M. McShane, Douglas G. Altman, Willi Sauerbrei, Identification of Clinically Useful Cancer Prognostic Factors: What Are We Missing?, JNCI: Journal of the National Cancer Institute, Volume 97, Issue 14, 20 July 2005, Pages 1023–1025, https://doi.org/10.1093/jnci/dji193
Close

Navbar Search Filter Mobile Enter search term Search

The number of cancer prognostic markers that have been validated as clinically useful is pitifully small, despite decades of effort and money invested in marker research ( 13 ) . For nearly all markers, the product has been a collection of studies that are difficult to interpret because of inconsistencies in conclusions or a lack of comparability. Small, underpowered studies; poor study design; varying and sometimes inappropriate statistical analyses; and differences in assay methods or endpoint definitions are but a few of the explanations that have been offered for this disappointing state of affairs ( 411 ) . Researchers attempting to conduct meta-analyses of prognostic marker studies encounter many difficulties ( 1214 ) . In this issue of the Journal, a meta-analysis by Kyzas et al. ( 15 ) of the tumor suppressor protein TP53 as a prognostic factor in head and neck cancer provides compelling empirical evidence that selective reporting biases are a major impediment to conducting meaningful meta-analyses of prognostic marker studies. These biases have serious implications, not only for meta-analyses but also for interpretation of the cancer prognostic literature as a whole.

Kyzas et al. demonstrate bias in the estimated association between TP53 status and mortality relating to a number of factors. The prognostic importance of TP53 decreased as the pool of included studies increased from published and indexed studies to any published studies, and then to the full set of all published and unpublished studies with retrievable data. In addition, when marker positivity definition and choice of clinical endpoint were retrospectively partly standardized across studies, there was a decrease in the overall estimated association compared with that obtained when each study used its preferred marker positivity criterion and endpoint.

The authors are appropriately cautious in acknowledging that confidence intervals for the estimates across the different analysis sets do overlap and that one cannot determine what might have been the effect of additional nonretrievable data on the observed trends. Nonetheless, we believe that this study provides the most compelling evidence yet that the published prognostic literature is a serious distortion of the truth. Even the most trusting and optimistic readers must feel pangs of suspicion when comparing, for example, the highly asymmetric scatterplot of study-specific association estimates observed for the published indexed studies [figure 3, A in Kyzas et al. ( 15 ) ] to the nearly symmetric scatter of estimates resulting when unpublished studies are included [figure 3, C in Kyzas et al. ( 15 ) ]. There is a clear message that studies showing stronger associations were more likely to be published in indexed journals, and there appeared to be a preference for reporting endpoints that produced the “best” results. Careful examination of the causes and discussion of potential remedies for these problems is warranted.

Most prognostic marker studies are conducted on retrospective collections of specimens. Archived specimen collections with good clinical and pathologic annotation and reliably collected clinical follow-up information are rare; many existing collections are not representative of well-defined populations and accompanying data may not be reliable. Frequently, study sample sizes are driven more by the number of specimens available than by any rigorous design criteria or statistical power considerations. With specimens in hand, a typical marker study is much easier to conduct than a randomized clinical trial. Often, institutional review requirements are minimal and formal protocols are rare; data collection may be informal and not stringently quality controlled. If a small marker study produces uninterpretable or equivocal results, an investigator may choose to abandon it and never publish the results. When a marker effect is found to be non–statistically significant in a small study, investigators, reviewers, and editors may choose to dismiss the study as underpowered. However, if the same study had produced a highly statistically significant result, it may be labeled as promising and be published. In contrast, a randomized clinical trial that had undergone institutional review, and for which considerable time and money had been invested for the study design, monitoring, and analysis, would be far more likely to be published, regardless of the findings. Yet, there is consistent evidence even for randomized clinical trials of selective reporting in relation to statistical significance ( 16 , 17 ) .

Compounding the problem of selective publication is the tendency for many endpoints to be examined within the same prognostic marker study and, if assay values are dichotomized, for many cutpoints to be considered ( 18 , 19 ) . The impact of this “overanalysis” is a further bias resulting in inflated estimates of association and increased chances of spurious findings. If marker studies were conducted according to prespecified protocols with clear designation of primary endpoints and assay scoring methods, these problems would be largely avoided. As in clinical trials, central parts of the analysis strategy should ideally be predefined. Additional explanatory analyses are highly welcome, but the character of the analyses should be stated in the paper and should influence the interpretation. We expect that few of the studies reviewed by Kyzas et al. were conducted according to a predefined protocol or had study sample sizes chosen to provide adequate power to detect a specified marker effect size. Most of the studies included fewer than 100 patients, a number that is obviously too small for a reasonable multivariable analysis. More complete and transparent reporting of marker studies would make it easier to distinguish carefully designed and analyzed studies from haphazardly designed and over-analyzed studies. The forthcoming REMARK guidelines ( 20 ) are an attempt to improve the reporting of tumor maker prognostic studies, but substantial improvement of the primary studies is a prerequisite.

The availability of individual patient data has been cited ( 14 ) as an important advantage for conducting meta-analyses of prognostic marker studies because it may permit reanalyses of data to accomplish standardization of marker values and endpoint definitions or to achieve comparability of patient populations. This approach leads to more interpretable overall estimates of association. It is shameful that, when primary investigators of 64 studies were contacted by Kyzas et al. in attempts to retrieve raw data, those for 22 studies were unable or unwilling to make their data available. Many journals, including the Journal of the National Cancer Institute , require that genomic data such as gene expression microarray data be made available as a condition of publication. Why should data availability be important only for high-dimensional marker data such as microarray data? The message is clear that, if data are not made available at the time of study completion and publication, it is unlikely that they will be retrievable later. Moreover, careful documentation should be viewed as essential for not only data variable definitions but, more generally, also for all aspects of the design, conduct, and analysis of the study. If the data are not properly documented at the time of collection, they may be uninterpretable several years later. Individual patient data from poorly designed, conducted, or documented studies will be of little value.

The tumor marker research community must come to the same realization that clinical trialists came to decades ago. If sound scientific principles of careful study design, adequate study size, scrupulous data collection and documentation, and appropriate analysis strategies are not adhered to, the field will flounder. Culture changes will be required. Stable and adequate funding will be required to have necessary personnel and infrastructure to collect, annotate, and maintain valuable specimen collections essential for high-quality retrospective studies. More importantly, the necessity of large, definitive prospective studies or prospectively planned meta-analyses for tumor marker research must be recognized.

References

(1)

Hayes DF, Bast RC, Desch CE, Fritsche H Jr, Kemeny NE, Jessup JM, et al. Tumor marker utility grading system: a framework to evaluate clinical utility of tumor markers.

J Natl Cancer Inst

1996

;

88

:

1456

–66.

(2)

Bast RC Jr, Ravdin P, Hayes DF, Bates S, Fritsche H Jr, Jessup JM, et al. for the American Society of Clinical Oncology Tumor Markers Expert Panel. 2000 update of recommendations for the use of tumor markers in breast and colorectal cancer: clinical practice guidelines of the American Society of Clinical Oncology.

J Clin Oncol

2001

;

19

:

1865

–78.

(3)

Schilsky RL, Taube SE. Introduction: tumor markers as clinical cancer tests—are we there yet?

Semin Oncol

2002

;

29

:

211

–2.

(4)

McGuire WL. Breast cancer prognostic factors: evaluation guidelines.

J Natl Cancer Inst

1991

;

83

:

154

–5.

(5)

Fielding LP, Fenoglio-Preiser CM, Freedman LS. The future of prognostic factors in outcome prediction for patients with cancer.

Cancer

1992

;

70

:

2367

–77.

(6)

Burke HB, Henson DE. Criteria for prognostic factors and for an enhanced prognostic system.

Cancer

1993

;

72

:

3131

–5.

(7)

Concato J, Feinstein AR, Holford TR. The risk of determining risk with multivariable models.

Ann Intern Med

1993

;

118

:

201

–10.

(8)

Gasparini G, Pozza F, Harris AL. Evaluating the potential usefulness of new prognostic and predictive indicators in node-negative breast cancer patients.

J Natl Cancer Inst

1993

;

85

:

1206

–19.

(9)

Simon R, Altman DG. Statistical aspects of prognostic factor studies in oncology.

Br J Cancer

1994

;

69

:

979

–85.

(10)

Gasparini G. Prognostic variables in node-negative and node-positive breast cancer.

Breast Cancer Res Treat

1998

;

52

:

321

–31.

(11)

Hall PA, Going JJ. Predicting the future: a critical appraisal of cancer prognosis studies.

Histopathology

1999

;

35

:

489

–94.

(12)

Altman DG. Systematic reviews of evaluations of prognostic variables. In: Egger M, Davey Smith G, Altman DG, eds. Systematic reviews in health care. Meta-analysis in context. 2nd ed. London (UK): BMJ Books;

2001

. p. 228–47.

(13)

Altman DG. Systematic reviews of evaluations of prognostic variables.

BMJ

2001

;

323

:

224

–8.

(14)

Riley RD, Abrams KR, Sutton AJ, Lambert PC, Jones DR, Heney D, et al. Reporting of prognostic markers: current problems and development of guidelines for evidence-based practice in the future.

Br J Cancer

2003

;

88

:

1191

–8.

(15)

Kyzas PA, Loizou KT, Ioannidis JPA. Selective reporting biases in cancer prognostic factor studies.

J Natl Cancer Inst

2005

;

97

:

1043

–55.

(16)

Sutton AJ, Abrams KR, Jones DR, Sheldon TA, Song F. Systematic reviews of trials and other studies.

Health Technol Assess

1998

;

2

:

1

–276.

(17)

Chan AW, Hrobjartsson A, Haahr MT, Gotzsche PC, Altman DG. Empirical evidence for selective reporting of outcomes in randomized trials: Comparison of protocols to published articles.

JAMA

2004

;

291

:

2457

–65.

(18)

Altman DG, De Stavola BL, Love SB, Stepniewska KA. Review of survival analyses published in cancer journals.

Br J Cancer

1995

;

72

:

511

–8.

(19)

Altman DG, Lausen B, Sauerbrei W, Schumacher M. Dangers of using “optimal” cutpoints in the evaluation of prognostic factors.

J Natl Cancer Inst

1994

;

86

:

829

–35.

(20)

McShane LM, Altman DG, Sauerbrei W, Taube SE, Gion M, Clark GM; for the Statistics Subcommittee of the NCI-EORTC Working Group on Cancer Diagnostics. Reporting recommendations for tumor marker prognostic studies (REMARK). J Natl Cancer Inst. In press

2005

.

© The Author 2005. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oupjournals.org.

Citations

Views

Altmetric

Metrics

Total Views 928

541 Pageviews

387 PDF Downloads

Since 2/1/2017

Month: Total Views:
February 2017 5
March 2017 2
April 2017 1
June 2017 4
July 2017 4
August 2017 3
September 2017 1
October 2017 5
November 2017 5
December 2017 35
January 2018 23
February 2018 5
March 2018 22
April 2018 11
May 2018 19
June 2018 12
July 2018 14
August 2018 14
September 2018 11
October 2018 13
November 2018 15
December 2018 12
January 2019 11
February 2019 10
March 2019 15
April 2019 22
May 2019 27
June 2019 13
July 2019 23
August 2019 10
September 2019 10
October 2019 16
November 2019 12
December 2019 10
January 2020 16
February 2020 15
March 2020 10
April 2020 12
May 2020 7
June 2020 10
July 2020 9
August 2020 5
September 2020 9
October 2020 11
November 2020 9
December 2020 11
January 2021 5
February 2021 11
March 2021 10
April 2021 6
May 2021 10
June 2021 5
July 2021 3
August 2021 1
September 2021 6
October 2021 6
November 2021 10
December 2021 2
January 2022 6
February 2022 9
April 2022 12
May 2022 3
June 2022 2
July 2022 11
August 2022 14
September 2022 22
October 2022 12
November 2022 12
December 2022 6
January 2023 7
February 2023 2
March 2023 6
April 2023 8
May 2023 3
June 2023 1
July 2023 6
August 2023 33
September 2023 8
October 2023 11
November 2023 5
December 2023 7
January 2024 8
February 2024 36
March 2024 10
April 2024 13
May 2024 3
June 2024 6
July 2024 10
August 2024 7
September 2024 13
October 2024 2

Citations

66 Web of Science

×

Email alerts

Citing articles via

More from Oxford Academic