Abstract
The use of targeted agents in the treatment of pediatric low-grade gliomas (pLGGs) relies on the determination of molecular status. It has been shown that genetic alterations in pLGG can be identified non-invasively using MRI-based radiomic features or convolutional neural networks (CNNs). We aimed to build and assess a combined radiomics and CNN non-invasive pLGG molecular status identification model. This retrospective study used the tumor regions, manually segmented from T2-FLAIR MR images, of 336 patients treated for pLGG between 1999 and 2018. We designed a CNN and Random Forest radiomics model, along with a model relying on a combination of CNN and radiomic features, to predict the genetic status of pLGG. Additionally, we investigated whether CNNs could predict radiomic feature values from MR images. The combined model (mean AUC: 0.824) outperformed the radiomics model (0.802) and CNN (0.764). The differences in model performance were statistically significant (p-values < 0.05). The CNN was able to learn predictive radiomic features such as surface-to-volume ratio (average correlation: 0.864), and difference matrix dependence non-uniformity normalized (0.924) well but was unable to learn others such as run-length matrix variance (− 0.017) and non-uniformity normalized (− 0.042). Our results show that a model relying on both CNN and radiomic-based features performs better than either approach separately in differentiating the genetic status of pLGGs, and that CNNs are unable to express all handcrafted features.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Introduction
Pediatric low-grade gliomas (pLGG) are a heterogeneous set of tumors with various histological features that can occur anywhere in the central nervous system1, and are sometimes referred to as low-grade neuroepithelial tumors. pLGGs account for approximately 40% of central nervous system tumors in children and are the most prevalent brain tumor during childhood2. With 10-year overall survival exceeding 90%3 death from pLGG is relatively rare. Nonetheless, with a progression-free survival rate of 50%, the morbidity of pLGG is high with many patients needing adjuvant therapy3,4.
Recent findings showed that pLGG typically upregulates the mitogen-activated protein kinase pathway, which led to the implementation of targeted therapeutics that can supplement or replace classic cytotoxic treatments3. The use of these targeted therapeutics requires accurate detection of the genetic alteration underlying pLGG, which is typically determined through tumor tissue acquired by surgery5. In some cases, such as in midline pLGG, it may not be feasible to obtain a biopsy due to the risk of neurologic compromise related to the surgery. Brain biopsies are costly, have numerous risks, leave residual tumor, and occasionally fail due to an insufficient sample.
Radiomics is the process of extracting quantitative features, which can be used to predict characteristics of patients, from radiological images6. Over the last decade, radiomics has emerged as a method of decoding tumor phenotypes non-invasively7,8,9. Radiogenomics, a research area within radiomics that combines “radiology” and “genomics”10, aims to identify relationships between imaging features and genomic data11. Numerous studies have shown the utility of radiomics and radiogenomics in neuro-oncology, including for predicting recurrence, survival, and genetic status12.
The two most common genetic alterations in pLGG are KIAA1549-BRAF fusion (BRAF Fusion) and BRAF p.V600E (BRAF Mutation)13. Wagner et al. showed that it is possible to differentiate between these two alterations using a machine learning (ML) approach relying on radiomic features extracted from T2-weighted fluid-attenuated inversion recovery (FLAIR) MRI sequences14,15. These results demonstrated the feasibility of a non-invasive pLGG genetic alteration identification model. Since then, multiple studies16,17,18 have shown that radiomics can be used to determine the BRAF status of all pLGGs, rather than just differentiating between the two most common genetic alterations. Tak et al.19, showed that using Convolutional neural networks (CNNs), an alternative approach to radiomics for extracting information from radiological images, it is possible to accurately sort pLGGs into three groups: BRAF Fusion, BRAF Mutation, and non-BRAF altered. Here we evaluated a more advanced approach, relying on both CNNs and radiomics, on the same classification task.
Radiomics relies on handcrafted features, while CNNs are trained to extract differentiative features (Fig. 1)20. The human design of radiomic features limits the amount of useful information that the radiomics approach can extract from medical images21. CNNs do not have this limitation, giving them greater expressive power22, thus they are often thought to be superior to radiomics21. However, CNNs have limitations of their own, namely, they need large datasets to learn from20,23,24. Nevertheless, CNNs have exploded in popularity25, and have become the dominant method to tackle a variety of medical imaging tasks26. The prominence of CNNs for medical imaging tasks in the literature suggests that it is commonly believed that radiomic features are redundant to those discovered by CNNs. According to Orlhac et al., this belief stems from the fact that theoretically, handcrafted features represent only a small subset of the features a neural network can capture22. However, in practice, neural networks may have trouble learning to represent certain handcrafted features due to restrictions such as limited data22.
Zhang et al. showed that radiomic and CNN features extracted from CT images of pancreatic ductal adenocarcinoma are only weakly related, suggesting a complementary relationship27. Other studies have performed more explicit tests and found that CNNs and radiomics work better together for various tasks including the classification of breast tumors28,29, ground glass nodules30, central nervous system tumors31, and adult gliomas32. We hypothesized that a combined approach would result in a more accurate pLGG genetic status classification model, thus we aimed to assess the ability of CNNs to complement handcrafted radiomic features. Additionally, we aimed to evaluate whether CNNs can capture all of the information contained in handcrafted features.
Materials and methods
Patients and data
All methods of this retrospective study were performed in accordance with the guidelines and regulations of the research ethics board of The Hospital for Sick Children (SickKids) (Toronto, Canada), which approved the study and waived informed consent. The electronic health record database at SickKids was screened for patients treated for pLGG between 1999 and 2018 with pre-therapeutic MR brain imaging and molecular characterization available. All MRIs used in this study were acquired prior to any treatment or intervention. We used FLAIR as our primary imaging sequence because it is useful for assessing both tumor volume and cysts, and was available for the most patients. Furthermore, FLAIR depicts the tumor and surrounding area better than contrast-enhanced T1-weighted images4, is known to be sensitive to leptomeningeal spread and non-contrast enhancing lesions5, and is considered to be the most sensitive technique to detect brain tumors overall6. Thus, of the 397 patients found through initial screening, 45 were omitted due to the absence of a FLAIR image, while 16 others were omitted because the FLAIR image was motion-degraded, leaving 336 patients to be included in this study. Many of these patients also had other sequences available: T2-weighted (265), T1-weighted with (270) and without (260) contrast. All available MRI sequences as well as patient molecular status, age, and sex were manually extracted from the electronic health record database. The demographics of the cohort are summarized in Table 1. Some of these patients (11514, 21515, 22033) were used in previous studies which only accounted for the two most common genetic alterations, resulting in less realistic and clinically useful models compared to this current study.
Molecular subtyping
A stepwise approach was used for the molecular characterization of pLGG, as previously described in14. First immunohistochemistry was used to detect BRAF mutation, then either a nCounter Metabolic Pathways Panel (NanoString Technologies) or fluorescence in situ hybridization was used to identify BRAF Fusion. Samples that were negative for BRAF Mutation and Fusion were analyzed further using other sequencing strategies detailed in13, such as RNA sequencing or panel DNA sequencing. For most patients, formalin-fixed paraffin-embedded tissue obtained during biopsy or resection was used for the molecular analysis, otherwise, frozen tissue was used.
MRI acquisition, segmentation, and preprocessing
Patients underwent brain MRI at 1.5 T or 3 T using either the Achieva, (Philips Healthcare), or Magnetom Skyra (Siemens Healthineers) system. MRI data were deidentified after being extracted from the PACS at SickKids. Each patient had a 2D FLAIR sequence, either axial or coronal.
As described in14, tumor regions were identified through segmentation of the FLAIR images by a pediatric neuroradiologist (MWW). Segmentations were validated by a senior neuroradiologist with more than 20 years of experience (BEW). The level tracing-effect tool of 3D Slicer34,35 (Version 4.10.2) was used to perform semi automated tumor segmentation. The tumor region alone was used as input to our machine-learning models. Images were z-score normalized, bias-corrected, and isotropically resampled to a resolution of 240 × 240 × 155 using 3D Slicer, to help account for differences in slice thickness, field strength, and pixel spacing.
Radiomic feature extraction
The same pipeline as in14 was used to generate radiomic features from the MRIs. The SlicerRadiomics extension of 3D Slicer was used to access PyRadiomics, an open-sourced package for radiomic feature extraction36. Default PyRadiomics settings were used, for example, bin width was set to 25. For each patient, 851 radiomic features, including first-order, shape, and second-order (texture) features, were extracted. The full list of radiomic features can be accessed in the Online Supplemental Data of14.
ML models
Each of our models was trained on a multiclass classification task with three class labels: BRAF Fusion, BRAF Mutation, and non-BRAF altered, a heterogeneous class containing numerous genetic alterations. The radiomics approach relied on a Random Forest (RF) model. We experimented with two different CNN architectures, a well-established deep neural network, the 3D ResNet37, and a custom shallow 3D CNN (Fig. 2) with three convolutional layers and two hidden fully connected layers. For the shallow CNN, the Leaky ReLU activation function and batch normalization were used between each layer, while max-pooling was employed after each convolutional layer. For the combined method we implemented feature-level fusion23 similar to the techniques implemented in27 and32, where CNN features are extracted from the fully-connected hidden layers and then used together with the radiomic features to train an RF (Fig. 3). We trained CNNs on both the FLAIR sequence alone and in conjunction with the other sequences (T2-weighted, T1-weighted with and without contrast), in an attempt to improve the accuracy of the model by including additional information. The only difference between these two configurations was the number of input channels (one versus four), and when training the model on four sequences we employed an input-level dropout approach38 to deal with missing sequences since not all patients had every sequence available.
In addition to testing whether CNNs and radiomics were complementary by combining them, we tested whether the CNNs could learn radiomics features as labels of MR images (Fig. 3). It was not computationally feasible to test the ability of the CNN to learn all 851 radiomic features, so we focused on the features that were the most predictive of pLGG molecular status in our radiomics model according to permutation importance. The rationale for using this subset of radiomic features was that if we found that the CNN could not learn them, we would gain some insight into the limitations of CNNs on this classification task. To the best of our knowledge, our experiment was the first to use real MRI images to test the ability of CNNs to learn radiomic features.
Along with the preprocessing steps outlined above that were performed to prepare the MRIs for radiomic feature extraction, each image was registered to the SRI24 atlas39 prior to being used with the CNN. Registration is commonly included in CNN pipelines, for instance, the Brain Tumor Segmentation challenge40,41,42 registers images to the SRI24 atlas. Image registration is thought to help CNNs identify positional information by aligning images such that each voxel represents the same anatomical location across all images. The radiomic approach does not account for positional information, which is not captured by typical radiomic features. Thus we did not include registration in the radiomics pipeline, since registration warps and resizes the brain, adding noise to the radiomic features, without any clear benefit.
Experiment configuration
To build the RF radiomics model we split the patients into test (20%) and development (80%) sets, where the development set is the training set plus the validation set, 25 times and performed five-fold cross-validation within each development set to tune hyperparameters (Table 2). Under this nested-cross validation approach, results were unlikely to be influenced by random data splitting. Due to the high computational load of training neural networks, a nested cross-validation scheme was not feasible for experiments using CNNs. We still repeated our experiments to avoid results biased by (un)lucky data splits by resampling test (20%) and development (80%) sets 25 times for the classification tasks and 10 times when attempting to learn radiomics with CNN. However, we only optimized CNN hyperparameters (Table 2) over a single split of the development set (still into five parts, four of which were used for training, while the last one was used for validation), rather than using five-fold cross-validation. Stratified sampling was used for all data splits, to account for class imbalances. When training CNNs the batch size was set to 16 images, and we used a cosine annealing period43 of 50 epochs, with a single warm restart for a total of 100 epochs of training. Kaiming Normal weight initialization44 was implemented to initialize weight parameters in both the convolutional and fully connected layers. Dropout was used during model training and switched off for inference. The final CNN used for evaluation on the test set was taken from the best epoch of the optimal set of hyperparameters, as measured by validation loss on the validation portion of the development set.
Both CNN and RF classification models were trained using cross-entropy loss and evaluated by One-Vs-Rest Area Under the Receiver Operating Characteristic Curve (AUC) on held-out test data. Correlation between predicted and true radiomic feature values was used to evaluate CNNs that were trained, using mean squared error loss, to learn radiomics features as MR image labels. The radiomics, CNN, and combined models were trained and tested on identical data splits. Resampling results in the test set of one trial containing samples from the training set of another trial, which invalidates the independence assumption of the traditional t-test45. To account for this dependence, we use the “corrected resampled t-test”46,47 to test for statistically significant differences in model performance. Python 3.11.0 was used to run all experiments. We relied upon Python’s Scikit-learn package 1.2.048 to execute ML concepts, while the PyTorch 1.13.0 library49 was used to implement deep-learning models.
Results
The performance of each of the models we trained is summarized in Table 3. On average, using the FLAIR sequence alone, the difference between the performance of the custom shallow CNN (mean AUC: 0.764) and the ResNet (0.758) was not statistically significant (p-value: 0.3836). Thus, we proceeded with the more computationally efficient custom shallow CNN, which is 10 × faster than the ResNet. The difference between the performance of the CNN when trained on FLAIR alone and with all four sequences (mean AUC: 0.757) was not significant (p-value: 0.272), so we continued with FLAIR alone for the remaining experiments.
The performance of the radiomics model (mean AUC: 0.802) eclipsed that of the CNN (0.764), but the combined model performed best (0.824). The differences were significant; p-values when testing whether the combined model was better than the radiomics model or the CNN alone were 0.0344 and 0.0002 respectively, while the p-value when comparing the radiomics model and the CNN was 0.0350.
The most important features in the radiomics model according to permutation importance are listed in Table 4. Figure 4 depicts the ability of the CNN to learn these features as labels from the FLAIR MR images. The CNN was able to learn gray level dependence matrix (GLDM) dependence non-uniformity normalized (average correlation: 0.924) and surface-to-volume ratio (0.864) well. It had more trouble, but some level of success, with gray level size zone matrix (GLSZM) zone percentage (0.735), flatness (0.563), and sphericity (0.532). The CNN had no ability to learn gray level run length matrix (GLRLM) non-uniformity normalized (− 0.042) or variance (− 0.017).
Discussion
This study explored the use of radiomics and CNNs to create a model that can non-invasively identify the underlying genetic alteration of pLGGs, labeling them as either BRAF Fusion, BRAF Mutation, or non-BRAF altered. We found that radiomics (mean AUC: 0.802) outperformed CNNs (0.764) and that a combined model performed better than either approach on its own (0.824). Our experiments also showed that utilizing a deeper model did not improve CNN performance. Furthermore, performance was similar with and without additional MRI sequences beyond FLAIR. We trained CNNs to learn predictive radiomic features from MR images and found that the CNN could estimate the value of certain radiomic features well, however, it was completely incapable of predicting others. These results suggest that the CNN may have performed worse than radiomics because it could not extract all the information contained in the radiomic features.
Handcrafted features have fallen out of favor for medical imaging tasks50, while neural network-based approaches have grown in prominence, in part because, theoretically, they can learn to discover any handcrafted feature from an image; but this does not always play out in practice. We think that it is too early to give up on the traditional ML techniques relying on handcrafted features. It is yet unclear as to whether the drawbacks of radiomics are more problematic than those of CNNs. There have been only a few studies that compare the two approaches directly, and the results have been inconsistent. CNNs have been shown to outperform radiomics for the identification of schizophrenia51, axillary lymph node metastasis21, and malignant breast lesions20. Radiomics surpassed CNNs for the classification of central nervous system tumors31, and the differentiation of malignant and benign ground glass nodules30. Our study adds another data point to this limited body of evidence and suggests that, contrary to popular belief, CNNs alone might not always be the best option for medical imaging tasks.
Klyuzhin et al.24 found that CNNs can learn first-order intensity and size-related radiomic features but are less able to learn shape irregularity and heterogeneity properties on a synthetic PET dataset. Their findings provided preliminary evidence that CNNs used alone in medical imaging are fundamentally limited due to their inability to capture radiomic features associated with clinical outcomes22. Our investigation, performed on MR images from real patients, provides further insights into the relationship between CNNs and radiomic features. In line with the results of24, we found that CNNs are better at learning features influenced by size, like surface-to-volume ratio, than shape-based features, like sphericity and flatness. Furthermore, we explored the ability of the CNN to learn texture features derived from the GLDM, GLRLM, and GLSZM. Klyuzhin et al. concluded that CNNs have a limited ability to learn radiomic texture features generally, whereas our results suggest that CNNs have trouble learning only certain pLGG radiomic texture features. GLDM dependence non-uniformity normalized was learned well, though the CNN performed worse on GLSZM zone percentage and had no ability to learn GLRLM variance or gray-level non-uniformity normalized. Overall, our results provide supporting evidence for the conclusion of Klyuzhin et al.24, that CNNs are not capable of capturing all of the tumor information contained in handcrafted radiomic features.
There are limitations to this study. First, though our dataset is large from a pediatric neuro-oncology perspective (336 patients), from an ML standpoint, it is relatively small and was collected from a single institution. Prospective data and data from other institutions are needed to evaluate the generalizability of our approach and the reliability of our model. Nevertheless, our dataset is diverse, having been collected over two decades using different MRI scanners and field strengths, and our results were found to be statistically significant under a rigorous nested cross-validation approach. Second, our claims about the performance of CNNs, radiomics, and combined models, the (in)ability of CNNs to learn certain radiomic features, and the lack of benefit when including different MRI sequences beyond FLAIR and larger CNNs, are limited to the specific scenarios we explored in this study. There are countless other combinations of radiomics models, neural networks, experiment configurations and preprocessing techniques, that could result in different conclusions; these need to be explored to evaluate the generalizability of our results. However, our results align with related studies24,27,28,29,30,31,32 that used different configurations, models, and data sources, giving us confidence that our conclusions will be confirmed by future studies. Third, our analysis was based on the entire tumor region. Performance may have been better with more specific labels (edema, necrosis, and enhancing vs non-enhancing structures); this exploration has been left to future work. Finally, it is not clear from our experiments why the CNN underperformed radiomics on the classification task and was unable to learn some radiomic features. Further experiments on larger datasets are required to determine whether CNN performance was limited because of a lack of data, or because there are inherent limitations to the performance of CNNs due to their design.
Conclusion
In this study, we created a model capable of non-invasively classifying pLGGs by BRAF status. We investigated the performance of radiomics and CNNs both separately and combined for this classification task, and found that the combined model relying on both CNNs and radiomic features performed best. Furthermore, we identified radiomic features that CNNs had trouble learning, uncovering limitations of CNNs in terms of the types of information they can extract from medical images. Future studies with large external and prospective datasets are necessary to improve diagnostic accuracy and further validate the robustness of the results.
Data availability
The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request pending the approval of the institution(s) and trial/study investigators who contributed to the dataset.
References
Sievert, A. J. & Fisher, M. J. Pediatric low-grade gliomas. J. Child Neurol. 24, 1397–1408 (2009).
Ostrom, Q. T. et al. CBTRUS statistical report: primary brain and other central nervous system tumors diagnosed in the United States in 2013–2017. Neuro Oncol. 22, iv1–iv96 (2020).
Ryall, S., Tabori, U. & Hawkins, C. Pediatric low-grade glioma in the era of molecular diagnostics. Acta Neuropathol. Commun. 8, 30 (2020).
Armstrong, G. T. et al. Survival and long-term health and cognitive outcomes after low-grade glioma. Neuro Oncol. 13, 223–234 (2011).
Bennett, J. et al. Canadian pediatric neuro-oncology standards of practice. Front. Oncol. 10, 593192 (2020).
Lambin, P. et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol 14, 749–762 (2017).
Khalvati, F., Zhang, Y., Wong, A. & Haider, M. A. Radiomics. Encycl. Biomed. Eng. 2, 597–603 (2019).
Aerts, H. J. W. The potential of radiomic-based phenotyping in precision medicine: A review. JAMA Oncol. 2, 1636–1642 (2016).
Aerts, H. J. W. L. et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 5, 1–9 (2014).
Shui, L. et al. The era of radiogenomics in precision medicine: An emerging approach to support diagnosis, treatment decisions, and prognostication in oncology. Front. Oncol. 10, 570465 (2020).
Lo Gullo, R., Daimiel, I., Morris, E. A. & Pinker, K. Combining molecular and imaging metrics in cancer: radiOgenomics. Insights Imaging 11, 1 (2020).
Ak, M. et al. Evolving role and translation of radiomics and radiogenomics in adult and pediatric neuro-oncology. AJNR Am. J. Neuroradiol. https://doi.org/10.3174/ajnr.A7297 (2021).
Ryall, S. et al. Integrated molecular and clinical analysis of 1,000 pediatric low-grade gliomas. Cancer Cell 37, 569-583.e5 (2020).
Wagner, M. W. et al. Radiomics of pediatric low-grade gliomas: Toward a pretherapeutic differentiation of BRAF-mutated and BRAF-fused tumors. AJNR Am. J. Neuroradiol. 42, 759–765 (2021).
Kudus, K. et al. Increased confidence of radiomics facilitating pretherapeutic differentiation of BRAF-altered pediatric low-grade glioma. Eur. Radiol. https://doi.org/10.1007/s00330-023-10267-1 (2023).
Vafaeikia, P. et al. MRI-based end-to-end pediatric low-grade glioma segmentation and classification. Can. Assoc. Radiol. J. 75, 153–160 (2024).
Xu, J. et al. Radiomics features based on MRI predict BRAF V600E mutation in pediatric low-grade gliomas: A non-invasive method for molecular diagnosis. Clin Neurol. Neurosurg. 222, 107478 (2022).
Liu, Z. et al. Radiomic features from multiparametric magnetic resonance imaging predict molecular subgroups of pediatric low-grade gliomas. BMC Cancer 23, 848 (2023).
Tak, D. et al. Noninvasive molecular subtyping of pediatric low-grade glioma with self-supervised transfer learning. Radiol. Artif. Intell. 6, e230333 (2024).
Truhn, D. et al. Radiomic versus convolutional neural networks analysis for classification of contrast-enhancing lesions at multiparametric breast MRI. Radiology 290, 290–297 (2019).
Sun, Q. et al. Deep learning versus radiomics for predicting axillary lymph node metastasis of breast cancer using ultrasound images: Don’t forget the peritumoral region. Front. Oncol. 10, 53 (2020).
Orlhac, F. et al. Radiomics in PET imaging: A practical guide for newcomers. PET Clin. 16, 597–612 (2021).
Afshar, P. et al. From handcrafted to deep-learning-based cancer radiomics: Challenges and opportunities. IEEE Signal Process. Mag. 36, 132–160 (2019).
Klyuzhin, I. S. et al. Testing the ability of convolutional neural networks to learn radiomic features. Comput. Methods Progr. Biomed 219, 106750 (2022).
Sarvamangala, D. R. & Kulkarni, R. V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 15, 1–22 (2022).
Yamashita, R., Nishio, M., Do, R. K. G. & Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 9, 611–629 (2018).
Zhang, Y. et al. Improving prognostic performance in resectable pancreatic ductal adenocarcinoma using radiomics and deep learning features fusion in CT images. Sci. Rep. 11, 1–11 (2021).
Whitney, H. M. et al. Comparison of breast MRI tumor classification using human-engineered radiomics, transfer learning from deep convolutional neural networks, and fusion methods. Proc. IEEE 108, 163–177 (2020).
Antropova, N., Huynh, B. Q. & Giger, M. L. A deep feature fusion methodology for breast cancer diagnosis demonstrated on three imaging modality datasets. Med. Phys. 44, 5162–5171 (2017).
Hu, X. et al. Computer-aided diagnosis of ground glass pulmonary nodule by fusing deep learning and radiomics features. Phys. Med. Biol. 66, 065015 (2021).
Yun, J. et al. Radiomic features and multilayer perceptron network classifier: A robust MRI classification strategy for distinguishing glioblastoma from primary central nervous system lymphoma. Sci. Rep. 9, 1–10 (2019).
Choi, Y. S. et al. Fully automated hybrid approach to predict the IDH mutation status of gliomas via deep learning and radiomics. Neuro Oncol. 23, 304–313 (2021).
Wagner, M. W. et al. Dataset size sensitivity analysis of machine learning classifiers to differentiate molecular markers of paediatric low-grade gliomas based on MRI. Oncol. Radiother. 16, 01–06 (2022).
Kikinis, R., Pieper, S. D. & Vosburgh, K. G. 3D Slicer: A Platform for Subject-Specific Image Analysis, Visualization, and Clinical Support. In Intraoperative Imaging and Image-Guided Therapy (ed. Jolesz, F. A.) 277–289 (Springer, 2014).
Fedorov, A. et al. 3D Slicer as an image computing platform for the quantitative imaging network. Magn. Reson. Imaging 30, 1323–1341 (2012).
van Griethuysen, J. J. M. et al. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 77, e104–e107 (2017).
Kataoka, H., Wakamiya, T., Hara, K., Satoh, Y. (2020) Would mega-scale datasets further enhance spatiotemporal 3D CNNs? ArXiv
Grøvik, E., Yi, D., Iv, M., et al (2021) Handling missing MRI sequences in deep learning segmentation of brain metastases: A multicenter study. npj Digital Medicine 4:1–7
Rohlfing, T., Zahr, N. M., Sullivan, E. V. & Pfefferbaum, A. The SRI24 multichannel atlas of normal adult human brain structure. Hum. Brain Map. 31, 798–819 (2010).
Bakas, S. et al. Advancing The Cancer Genome Atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 4, 1–13 (2017).
Baid, U., Ghodasara, S., Bilello, M., et al (2021) The RSNA-ASNR-MICCAI BraTS 2021 Benchmark on brain tumor segmentation and radiogenomic classification. ArXiv
Menze, B. H. et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 34, 1993–2024 (2015).
Loshchilov, I., Hutter, F. (2017) SGDR: Stochastic gradient descent with warm restarts. International Conference on Learning Representations
He, K., Zhang, X., Ren, S., Sun, J. (2015) Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In: 2015 IEEE International Conference on Computer Vision (ICCV). pp 1026–1034
Dietterich, T. G. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 10, 1895–1923 (1998).
Bouckaert, R. R. & Frank, E. Evaluating the Replicability of Significance Tests for Comparing Learning Algorithms. Advances in Knowledge Discovery and Data Mining (Springer, 2004).
Nadeau, C. & Bengio, Y. Inference for the generalization error. Mach. Learn. 52, 239–281 (2003).
Pedregosa, F. et al. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 12, 2825–2830 (2011).
Paszke, A., Gross, S., Massa, F., et al (2019) PyTorch: an imperative style, high-performance deep learning library. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, USA, pp 8026–8037
Dutta, P., Upadhyay, P., De, M., Khalkar, RG. (2020) Medical image analysis using deep convolutional neural networks: CNN architectures and transfer learning. In: 2020 International Conference on Inventive Computation Technologies (ICICT). pp 175–180
Hu, M., Sim, K., Zhou, JH., et al (2020) Brain MRI-based 3D convolutional neural networks for classification of schizophrenia and controls. arXiv [cs.CV]
Acknowledgements
This research has been made possible with the financial support of the Canadian Institutes of Health Research (CIHR) (Funding Reference Number: 184015).
Author information
Authors and Affiliations
Contributions
K.K., B.E.W., and F.K. contributed to the design of the concept and study. K.K. contributed to the implementation of machine learning modules and running the experiments. K.K. and K.N. contributed to preprocessing of imaging data. J.B., L.N., U.T., and C.H. contributed to collecting, reviewing, and providing clinical data and genetic markers. M.W. and B.E.W. contributed to reviewing the imaging data and providing tumor segmentation for the data. K.K. wrote the first draft of the manuscript and all authors contributed to the reviewing and editing of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kudus, K., Wagner, M., Namdar, K. et al. Beyond hand-crafted features for pretherapeutic molecular status identification of pediatric low-grade gliomas. Sci Rep 14, 19102 (2024). https://doi.org/10.1038/s41598-024-69870-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-69870-x
- Springer Nature Limited