Discrepancies of assessments in a RECIST 1.1 phase II clinical trial - association between adjudication rate and variability in images and tumors selection

Hubert Beaumont, Tracey L Evans, Catherine Klifa, Ali Guermazi, Sae Rom Hong, Mustapha Chadjaa, Zsuzsanna Monostori, Hubert Beaumont, Tracey L Evans, Catherine Klifa, Ali Guermazi, Sae Rom Hong, Mustapha Chadjaa, Zsuzsanna Monostori

Abstract

Background: In imaging-based clinical trials, it is common practice to perform double reads for each image, discrepant interpretations can result from these two different evaluations. In this study we analyzed discrepancies that occurred between local investigators (LI) and blinded independent central review (BICR) by comparing reader-selected imaging scans and lesions. Our goal was to identify the causes of discrepant declarations of progressive disease (PD) between LI and BICR in a clinical trial.

Methods: We retrospectively analyzed imaging data from a RECIST 1.1-based, multi-sites, phase II clinical trial of 179 patients with adult small cell lung cancer, treated with Cabazitaxel compared to Topotecan. Any discrepancies in the determination of PD between LI and BICR readers were reviewed by a third-party adjudicator. For each imaging time point and reader, we recorded the selected target lesions, non-target lesions, and new lesions. Odds ratios were calculated to measure the association between discrepant declarations of PD and the differences in reviewed imaging scans (e.g. same imaging modality but with different reconstruction parameters) and selected lesions. Reasons for discrepancies were analyzed.

Results: The average number of target lesions found by LI and BICR was respectively 2.9 and 3.4 per patient (p < 0.05), 18.4% of these target lesions were actually non-measurable. LI and BICR performed their evaluations based on different baseline imaging scans for 59% of the patients, they selected at least one different target lesion in 85% of patients. A total of 36.7% of patients required adjudication. Reasons of adjudication included differences in 1) reporting new lesions (53.7%), 2) the measured change of the tumor burden (18.5%), and 3) the progression of non-target lesions (11.2%). The rate of discrepancy was not associated with the selection of non-measurable target lesions or with the readers' assessment of different images. Paradoxically, more discrepancies occurred when LI and BICR selected exactly the same target lesions at baseline compared to when readers selected not exactly the same lesions.

Conclusions: For a large proportion of evaluations, LI and BICR did not select the same imaging scans and target lesions but with a limited impact on the rate of discrepancy. The majority of discrepancies were explained by the difference in detecting new lesions.

Trial registration: ARD12166 ( https://ichgcp.net/clinical-trials-registry/NCT01500720 ).

Keywords: Inter-observer variability; Phase II; Response evaluation criteria in solid tumors; Small cell lung carcinoma; Tumor imaging.

Conflict of interest statement

Ethics approval and consent to participate

This study was exempted of ethic approval by the IRB and written informed consent was obtained from all subjects (patients) in this study.

Consent for publication

The final version of the manuscript has been reviewed and approved by all co-authors.

Competing interests

Hubert Beaumont and Catherine Klifa, as co-authors of this manuscript, declare relationships with the following companies Median Technologies.

Mustapha Chadjaa, declare relationships with the following companies SANOFI.

All other authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Figures

Fig. 1
Fig. 1
Flowchart of the ARD12166 trial. Using a common database of imaging scans, LI and BICR performed RECIST 1.1 evaluations. In case of evaluations discrepancies, an adjudicator was solicited to perform a third evaluations blinded from previous assessments
Fig. 2
Fig. 2
Example of one non-measurable lesion. Both readers targeted the same region of interest, but their perception of tumors’ boundary was. One reader (right) considered the hepatic tumor as coalescent, while the other (left) considered two distinct tumors. The inter-reader proportional difference was 73%. It should be noted that readers enabled different window levels
Fig. 3
Fig. 3
Number of selected target lesions. Number of selected target lesions by LI (blue bars) versus BICR (red bars) readers. Data are ordered first according to the higher number selected by BICR, and then by the higher number selected by the LI. The number of target lesions selected by BICR was significantly higher than that by the LI (pVal

Fig. 4

Anatomical sites where target lesions…

Fig. 4

Anatomical sites where target lesions have been selected. Pie chart displaying the proportion…

Fig. 4
Anatomical sites where target lesions have been selected. Pie chart displaying the proportion of sites where target lesions have been selected. Orange: pulmonary lesions; grey: hepatic lesions; yellow: nodal lesions; light blue: adrenal; navy blue: brain; green: undefined. Undefined lesions are in bone, spleen, pancreas, muscle, and kidney. Left: Location of LI target lesion selection. Right: Location of BICR target lesion selection
Fig. 4
Fig. 4
Anatomical sites where target lesions have been selected. Pie chart displaying the proportion of sites where target lesions have been selected. Orange: pulmonary lesions; grey: hepatic lesions; yellow: nodal lesions; light blue: adrenal; navy blue: brain; green: undefined. Undefined lesions are in bone, spleen, pancreas, muscle, and kidney. Left: Location of LI target lesion selection. Right: Location of BICR target lesion selection

References

    1. Food US, Document DAFDAB. Oncology drug advisory committee meeting BLA STN 125085/91. Washington, DC: Avastin® (bevacizumab). US Food Drug Adm; 2007.
    1. Tang P A, Pond GR, Chen EX. influence of an independent review committee on assessment of response rate and progression-free survival in phase III clinical trials. Ann Oncol 2010;21:19–26.
    1. Zhao B, James LP, Moskowitz CS, Guo P, Ginsberg MS, Lefkowitz RA, et al. Evaluating Variability in Tumor Measurements from Same-day Repeat CT Scans of Patients with Non–Small Cell Lung Cancer 1. Radiology [Internet]. 2009 [cited 2014 Jan 15];252:263–272. Available from:
    1. Yoon SH, Kim KW, Goo JM, Kim D-W, Hahn S. Observer variability in RECIST-based tumour burden measurements: a meta-analysis. Eur J Cancer [Internet]. 2016;53:5–15. Available from:
    1. Clarke LP, Sriram RD, Schilling LB. Imaging as a biomarker: standards for change measurements in therapy workshop summary. Acad Radiol [Internet] 2008;15:501–530. doi: 10.1016/j.acra.2007.10.021.
    1. Sridhara R, Mandrekar SJ, Dodd LE. Missing Data and Measurement Variability in Assessing Progression-Free Survival Endpoint in Randomized Clinical Trials. Clin Cancer Res [Internet]. 2013;19:2613–2620. Available from:
    1. Ford R, Schwartz L, Dancey J, Dodd LEE, Eisenhauer EA , Gwyther S, et al. Lessons learned from independent central review. Eur J Cancer [internet]. Elsevier Ltd; 2009;45:268–274. Available from: 10.1016/j.ejca.2008.10.031
    1. Levy MA, Rubin DL. Tool support to enable evaluation of the clinical response to treatment. AMIA Annu Symp Proc. 2008:399–403.
    1. Cohen KL, Gönen M, Ford RR. Monitoring reader metrics in blinded independent central review of oncology studies. J Clin Trials. 2015;5:1–6.
    1. Ford R, O’ Neal M, Moskowitz S, Fraunberger J. Adjudication Rates between Readers in Blinded Independent Central Review of Oncology Studies. J Clin Trials [Internet]. 2016;06. Available from:
    1. Akkerhuis KMK, Deckers JW, Boersma E, Harrington RA, Stepinska J, Mahaffey KW, et al. Geographic variability in outcomes within an international trial of glycoprotein IIb/IIIa inhibition in patients with acute coronary syndromes. Results from PURSUIT. Eur Heart J. 2000;21:371–381. doi: 10.1053/euhj.1999.1743.
    1. Skougaard K, McCullagh MJD, Nielsen D, Hendel HW, Jensen BV, Johannesen HH. Observer variability in a phase II trial - assessing consistency in RECIST application. Acta Oncol [Internet]. 2012 [cited 2014 Apr 30];51:774–780. Available from:
    1. Eisenhauer EA, Therasse P, Bogaerts J, Schwartz LH, Sargent D, Ford R, et al. New response evaluation criteria in solid tumours: revised RECIST guideline (version 1.1). Eur J Cancer [internet]. Elsevier Ltd; 2009;45:228–247. Available from: 10.1016/j.ejca.2008.10.026
    1. Zhao B, Tan Y, Bell D, Marley S. Intra-and inter-reader variability in uni-dimensional, bi-dimensional, and volumetric measurements of solid tumors on CT scans reconstructed at different slice intervals. Eur J … [Internet]. 2013 [cited 2014 Jun 20];82:1–22. Available from:
    1. Schwartz LH, Mazumdar M, Brown W, Smith A, Panicek DM. Variability in response assessment in solid tumors: effect of number of lesions chosen for measurement. Clin Cancer Res. 2003;9:4318–4323.
    1. Kim EJ, Kim SR, Gang SJ, Park SY, Han YM. Tonsillar metastasis of small cell lung Cancer in a patient with idiopathic pulmonary fibrosis. Medicine (Baltimore) 2015;94:1–4.
    1. Team RDC. R: A Language and Environment for Statistical Computing [Internet]. Team RDC, editor. R Found. Stat. Comput. R Foundation for Statistical Computing; 2011. p. 409. Available from:
    1. Borradaile K, Ford R. Analysis of the rate of non-target disease progression in patients with stable or responding target disease by the response evaluation criteria in solid tumors (RECIST). J Clin Oncol (meeting Abstr) [internet]. Proc Am Soc Clin Oncol. 2009:27 Available from: .
    1. Borradaile K, Ford R, O’Neal M, Byrne K. Analysis of the cause of discordance between two radiologists in the assessment of radiographic response and progression for subjects enrolled in breast Cancer clinical trials employing blinded independent central review. J Clin Oncol (meeting Abstr [internet]). Chicago, IL. Proc Am Soc Clin Oncol. 2010:28 Available from: .
    1. Schwartz LH, Bogaerts J, Ford R, Shankar L, Therasse P, Gwyther S, et al. Evaluation of lymph nodes with RECIST 1.1. Eur J Cancer [Internet]. Elsevier Ltd; 2009;45:261–267. Available from: 10.1016/j.ejca.2008.10.028
    1. Mozley D. Assessing the measurement variability of lung lesions in patient Data Sets (1B Study)—A Researcher’s Perspective: the Problem of Intra-reader Variability. QIBA Newsl [Internet]. 2013;5. Available from: .
    1. Keil S, Barabasch A, Dirrichs T, Bruners P, Hansen NL, Bieling HB, et al. Target lesion selection: an important factor causing variability of response classification in the Response Evaluation Criteria for Solid Tumors 1.1. Invest Radiol [Internet]. 2014;49:509–517. Available from:
    1. Mitchel JT, Kim YJ, Choi J, Park G, Cappi S, Horn D, et al. Evaluation of Data Entry Errors and Data Changes to an Electronic Data Capture Clinical Trial Database. Drug Inf J [Internet]. 2011;45:421–430. Available from:
    1. Karmakar A, Kumtakar A, Sehgal H, Kumar S, Kalyanpur A. Interobserver variation in response evaluation criteria in solid tumors 1. 1. Acad Radiol [Internet]. 2018; Available from: .
    1. Machado L, Apolo A, Steinberg S, Folio L. Radiology reports with hyperlinks improve target lesion selection and measurement concordance in Cancer trials. Am J Roentgenol. 2017;208:W31–W37. doi: 10.2214/AJR.16.16845.
    1. Mozley PD, Bendtsen C, Zhao B, Schwartz LH, Thorn M, Rong Y, et al. Measurement of tumor volumes improves RECIST-based response assessments in advanced lung cancer. Transl Oncol [Internet]. Neoplasia Press, Inc.; 2012 [cited 2014 Jan 15];5:19–25. Available from:
    1. Dinkel J, Khalilzadeh O, Hintze C, Fabel M, Puderbach M, Eichinger M, et al. Inter-observer reproducibility of semi-automatic tumor diameter measurement and volumetric analysis in patients with lung cancer. Lung Cancer [Internet]. 2013;82:76–82. Available from:
    1. Nodine CF, Mello-Thoms C. The role of expertise in radiologic image interpretation. In: Samei E, Krupinski EA, editors. Handb med image percept tech: Cambridge University Press; 2010. p. 139–56.

Source: PubMed

3
Subskrybuj