Interinstitutional Portability of a Deep Learning Brain MRI Lesion Segmentation Algorithm - PubMed (original) (raw)
Interinstitutional Portability of a Deep Learning Brain MRI Lesion Segmentation Algorithm
Andreas M Rauschecker et al. Radiol Artif Intell. 2021.
Abstract
Purpose: To assess how well a brain MRI lesion segmentation algorithm trained at one institution performed at another institution, and to assess the effect of multi-institutional training datasets for mitigating performance loss.
Materials and methods: In this retrospective study, a three-dimensional U-Net for brain MRI abnormality segmentation was trained on data from 293 patients from one institution (IN1) (median age, 54 years; 165 women; patients treated between 2008 and 2018) and tested on data from 51 patients from a second institution (IN2) (median age, 46 years; 27 women; patients treated between 2003 and 2019). The model was then trained on additional data from various sources: (a) 285 multi-institution brain tumor segmentations, (b) 198 IN2 brain tumor segmentations, and (c) 34 IN2 lesion segmentations from various brain pathologic conditions. All trained models were tested on IN1 and external IN2 test datasets, assessing segmentation performance using Dice coefficients.
Results: The U-Net accurately segmented brain MRI lesions across various pathologic conditions. Performance was lower when tested at an external institution (median Dice score, 0.70 [IN2] vs 0.76 [IN1]). Addition of 483 training cases of a single pathologic condition, including from IN2, did not raise performance (median Dice score, 0.72; P = .10). Addition of IN2 training data with heterogeneous pathologic features, representing only 10% (34 of 329) of total training data, increased performance to baseline (Dice score, 0.77; P < .001). This final model produced total lesion volumes with a high correlation to the reference standard (Spearman r = 0.98).
Conclusion: For brain MRI lesion segmentation, adding a modest amount of relevant training data from an external institution to a previously trained model supported successful application of the model to this external institution.Keywords: Neural Networks, Brain/Brain Stem, Segmentation Supplemental material is available for this article. © RSNA, 2021.
Keywords: Brain/Brain Stem; Neural Networks; Segmentation.
2021 by the Radiological Society of North America, Inc.
Conflict of interest statement
Disclosures of conflicts of interest: A.M.R. Institution received American Society of Neuroradiology (ASNR) trainee grant and Carestream Health/Radiological Society of North America (RSNA) research scholar grant; former member of Radiology: Artificial Intelligence trainee editorial board. T.J.G. No relevant relationships. P.N. Formerly employed by Perceus and PrinterPress (urologic devices and orthopedic devices); grants/grants pending from Perceus for a urology device; patent from Perceus for a urology device; stock/stock options in PrinterPress (orthopedic devices). M.T.D. No relevant relationships. D.A.W. Consultancy fee from Galileo CDS for work not related to this publication; stock/stock options in Galileo CDS received for work not related to this publication. E.C. No relevant relationships. J.B.C. No relevant relationships. L.P.S. No relevant relationships. J.D.R. Institution received ASNR neuroradiology research grant in AI; former member of Radiology: Artificial Intelligence trainee editorial board. C.P.H. Medical imaging consultant for GE Healthcare; research travel expenses from Siemens Healthineers; member of data monitoring and safety boards for Insightec and UniQure; former member of Radiology editorial board.
Figures
Figure 1:
Flow diagram demonstrates use of data from various institutions for training and testing. (A) Data from the 51 patients from institution 2 (IN2) formed the primary test set on which the previously trained model was tested. This dataset was then divided into three subsets of training data (n = 34) and independent test data (n = 17), forming a threefold cross-validation test sample. (B) The original model (M1) was trained on 293 patients’ data from one institution (IN1) and tested on both the same institution’s independent test data (replication of Duong et al [6]) and on IN2 data. Additional experiments were performed using training data composed of other data sources combined with the original dataset (BraTS [M1+B], IN2 brain tumor data [M1+B+2T], and IN2 training data [M1+2]). Fine-tuning with IN2 data (MFT (1+2)) was also explored in another set of experiments. BraTS = Multimodal Brain Tumor Segmentation challenge, M1 = model trained on IN1 patient data, M1+B = model trained on IN1 patient data + BraTS, M1+B+2T = model trained on IN1 patient data + BraTS + data from IN2 patients with tumors, M1+2 = model trained on IN1 + IN2 data, model MFT (1+2) = trained on IN1 with sequential fine-tuning with IN2, M2 = model trained on IN2 patient data.
Figure 2:
Segmentation accuracy (Dice scores) for five different versions of the convolutional neural network, with individual Dice scores for each patient indicated by a data point and median Dice scores indicated by horizontal black bars. The architecture and hyperparameters of the model remained identical, but training and testing cases varied. The horizontal dashed lined demonstrates baseline performance of the model trained at one institution and applied to the same institution, with a median Dice score (0.76) for comparison to the four other models, which used various mixtures of interinstitutional training data. IN1 = institution 1, IN2 = institution 2, M1 = model trained on IN1 patient data, M2 = model trained on IN2 data, M1+B= model trained on IN1 patient data + Multimodal Brain Tumor Segmentation challenge data, M1+2 = trained on IN1 + IN2 data.
Figure 3:
Representative single-section fluid-attenuated inversion recovery (FLAIR) images from 17 test samples using the final model (M1+2), which was trained using data from both institutions. Note the large variation in number, size, and extent of lesions. On the left of each pair is the original image, and on the right of each pair is the automated segmentation overlaid on the original FLAIR image. ADEM = acute disseminated encephalomyelitis, CNS = central nervous system, CADASIL = cerebral autosomal dominant arteriopathy with subcortical infarcts and leukoencephalopathy, MS = multiple sclerosis, M1+2 = model trained on institution 1 + institution 2 data, PML = progressive multifocal leukoencephalopathy, PRES = posterior reversible encephalopathy syndrome.
Figure 4:
Effect of adding institution 1 (IN1) training data on IN1 and institution 2 (IN2) test data performance. (A) Dice scores on IN1 (left panel) and IN2 (right panel) test data for models trained with various datasets, including different amounts of IN1 training data. Median Dice scores are indicated by horizontal bars. A small amount of IN1 training data (blue shading) was sufficient for high performance when tested on IN1 test data, while a small amount of IN2 training data (M2 or M1+2) improved performance on IN2 test data. (B) IN2 test data performance for various models with incremental addition of IN1 training data (green shading) to a baseline M2 model (yellow). Addition of IN1 training data had no effect on IN2-tested performance, while a small amount of IN2 data improved performance over the M1 baseline model (*** = P < .001,P = .06 where indicated). Statistical comparisons are to the bars in each panel indicated by “|” using Wilcoxon signed-rank test. M1 = trained on IN1 patient data, M2 = trained on IN2 data, M1+2 = trained on IN1 + IN2 data.
Figure 5:
Lesion volume estimates and volume effect on Dice score.(A) The correlation between true total lesion volume and predicted total lesion volume for M1+2 model (white squares) was high (Spearman r = 0.98). Black circles represent data from M1 tested on institution 1 (IN1) data as a reference. (B) Bland-Altman plot demonstrating the difference between predicted and true total lesion volume on the test set as a function of true total lesion volume in the range 0 to 500 cm3. Lines represent the mean (solid line) and ± 1.96 standard deviations (SD; dashed line) for M1 with IN1 test (black) and M1+2 with institution 2 (IN2) test (gray) models. Lesions larger than three SD above the mean training data lesion volume (n = 2) are excluded from this plot because they represent clear outliers. (C) Box plot of Dice score as a function of true total lesion volume for IN1-trained, IN2–fine-tuned model (MFT (1+2)). (*) indicates_P_ < .05, (**) indicates_P_ < .01. M1 = trained on IN1 patient data, M1+2 = trained on IN1 + IN2 data.
Figure 6:
Lesion segmentation performance of the bi-institutionally trained model according to various imaging parameters. (A) Dice score based on institution 2 (IN2) test data as a function of MRI machine model, grouped by MRI machine manufacturer. Dashed gray line represents median Dice score for the institution 1 (IN1)–trained, IN2–fine-tuned model (MFT (1+2)) tested on IN2 (0.78). There were no significant differences in median Dice score across 10 machine models (P = .66) or across three manufacturers (P = .71). (B) Dice scores in test data according to the number of training cases from the same machine model. (C) Dice scores as a function of field strength (left) and acquisition dimension (right). There was no effect of field strength (P = .45) or acquisition dimension (P = .69). Horizontal bars represent median Dice scores. All data shown are for MFT (1+2) tested on IN2 data using threefold cross-validation to ensure independent training and testing data. # = number of, 3D = three dimensional, 2D = two dimensional.
References
- Onofrey JA, Casetti-Dinescu DI, Lauritzen AD, et al. Generalizable Multi-Site Training and Testing Of Deep Neural Networks Using Image Normalization. In: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019),Conference Location, Conference date., Piscataway, NJ: IEEE, 2019; 348–351 10.1109/ISBI.2019.8759295. -DOI -PMC -PubMed
- Bhuva AN , Bai W , Lau C , et al . A Multicenter, Scan-Rescan, Human and Machine Learning CMR Study to Test Generalizability and Precision in Imaging Biomarker Analysis . Circ Cardiovasc Imaging 2019. ; 12 ( 10 ): e009214 . -PubMed
](