Next Article in Journal
CI-UNet: Application of Segmentation of Medical Images of the Human Torso
Previous Article in Journal
Assessment of the Fungistatic Properties of Calendula officinalis L. Water Extract and the Effect of Its Addition on the Quality of Wheat Bread
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reproducibility in Radiomics: A Comparison of Feature Extraction Methods and Two Independent Datasets

by
Hannah Mary T. Thomas
1,†,
Helen Y. C. Wang
2,3,†,
Amal Joseph Varghese
1,
Ellen M. Donovan
2,
Chris P. South
3,
Helen Saxby
4,
Andrew Nisbet
5,
Vineet Prakash
4,
Balu Krishna Sasidharan
1,
Simon Pradeep Pavamani
1,
Devakumar Devadhas
6,
Manu Mathew
1,
Rajesh Gunasingam Isiah
1 and
Philip M. Evans
2,*
1
Department of Radiation Oncology, Christian Medical College Vellore, Vellore 632004, Tamil Nadu, India
2
Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford GU2 7XH, UK
3
Department of Medical Physics, Royal Surrey NHS Foundation Trust, Guildford GU2 7XX, UK
4
St Luke’s Cancer Centre, Royal Surrey NHS Foundation Trust, Guildford GU2 7XX, UK
5
Department of Medical Physics and Biomedical Engineering, University College London, London WC1E 6BT, UK
6
Department of Nuclear Medicine, Christian Medical College Vellore, Vellore 632004, Tamil Nadu, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(12), 7291; https://doi.org/10.3390/app13127291
Submission received: 26 April 2023 / Revised: 12 June 2023 / Accepted: 16 June 2023 / Published: 19 June 2023
(This article belongs to the Section Applied Biosciences and Bioengineering)

Abstract

:

Featured Application

The application of this work is in radiomics for medical imaging analysis. It addresses the question of how to establish if radiomic features are stable and reproducible.

Abstract

Radiomics involves the extraction of information from medical images that are not visible to the human eye. There is evidence that these features can be used for treatment stratification and outcome prediction. However, there is much discussion about the reproducibility of results between different studies. This paper studies the reproducibility of CT texture features used in radiomics, comparing two feature extraction implementations, namely the MATLAB toolkit and Pyradiomics, when applied to independent datasets of CT scans of patients: (i) the open access RIDER dataset containing a set of repeat CT scans taken 15 min apart for 31 patients (RIDER Scan 1 and Scan 2, respectively) treated for lung cancer; and (ii) the open access HN1 dataset containing 137 patients treated for head and neck cancer. Gross tumor volume (GTV), manually outlined by an experienced observer available on both datasets, was used. The 43 common radiomics features available in MATLAB and Pyradiomics were calculated using two intensity-level quantization methods with and without an intensity threshold. Cases were ranked for each feature for all combinations of quantization parameters, and the Spearman’s rank coefficient, rs, calculated. Reproducibility was defined when a highly correlated feature in the RIDER dataset also correlated highly in the HN1 dataset, and vice versa. A total of 29 out of the 43 reported stable features were found to be highly reproducible between MATLAB and Pyradiomics implementations, having a consistently high correlation in rank ordering for RIDER Scan 1 and RIDER Scan 2 (rs > 0.8). 18/43 reported features were common in the RIDER and HN1 datasets, suggesting they may be agnostic to disease site. Useful radiomics features should be selected based on reproducibility. This study identified a set of features that meet this requirement and validated the methodology for evaluating reproducibility between datasets.

1. Introduction

There is growing evidence that standard-of-care medical images obtained from modalities such as CT, MRI, and PET contain more information than is visible to the human eye [1]. The high-throughput extraction and processing of the underlying information from radiological images is known as “radiomics”. The quantitative data obtained (imaging biomarkers) could potentially be used alongside the current gold standard of tumor evaluation and staging tools, including TNM staging [2], to aid clinical decision making such as personalized treatment planning.
The predictive power of radiomic features is dependent on having a large set of data. However, due to the nature of medical images, the size of the studies is often relatively small and based on a single dataset, restricting the impact of the results. To find candidates for reproducible biomarkers from the hundreds of features available from first, second, and higher-order statistical features of images, it is necessary for researchers to validate the results published by other groups [3]. This should be carried out using a separate dataset from the original study and considered a retrospective investigation. However, at least 50% of published studies have been described as poorly reported with incomplete methodologies and results for successful validation when an analysis of biomedical research was performed by Chalmers and Glasziou [3]. The precise cause of this serious lack of reproducibility in validation is unclear. The lack of standards for validating results, incomplete reporting of methodologies and results, and unrecognized confounding variables in the dataset used could all be to blame.
A recent systematic review of full-text articles in PubMed published in 2018 primarily addressed non-small cell lung cancer (NSCLC) and oropharyngeal cancer [4] and found only 7 out of 41 studies reported every methodology used in image acquisition, pre-processing, and feature extraction in detail. Out of 21 studies on NSCLC, 4 studies using CT images [5,6,7,8] and 1 study using PET (Positron Emission Tomography) images reported every methodologic aspect. The results identified the sensitivity of radiomic features in terms of repeatability and reproducibility to processing details such as the settings used in image acquisition, the image reconstruction algorithm, image preprocessing, and the software used to extract radiomic features. First-order features were reported to be more reproducible than shape metrics and texture features.
Our previously published study [9] analyzed radiomic features extracted from the CT component of PET/CT scans of patients with NSCLC, treated at the Royal Surrey NHS Foundation Trust (RSFT). The radiomics features were calculated using the toolkit of Vallières et al. [10], Which is available in the MATLAB package. This study found that a set of radiomics features were stable to settings used in image acquisition and reconstruction algorithms used in different scanner models. Features were also stable to variations in tumor delineation. However, features were sensitive to intensity quantization parameters, including (i) the number of intensity levels, (ii) the method of quantization to select the intensity levels, and (iii) the use of an intensity threshold around the tumor or organ being analyzed. These results show that different parameter choices in different datasets may help explain the results in the mentioned review papers [4,5,11]. Therefore, a question is: would these features be successfully reproduced and validated under different conditions, such as with a different lung cancer dataset, a different disease site, or using a different implementation of radiomics feature extraction?

2. Materials and Methods

The purpose of this paper is to investigate the generalizability of the findings from the initial study [9] and if a common set of CT radiomics features is stable. This was achieved first by evaluating which radiomics features are stable from the originally used 43 features of the Vallières toolkit, for a publicly available lung cancer dataset: the Reference Image Database to Evaluate Therapy Response (RIDER) [12]. As Pyradiomics [1] is one of the most used radiomics toolkits and provides the 43 features of the Vallières [10] plus 59 other original features, the RIDER dataset was also evaluated using Pyradiomics [13], and the results of the two toolkits were compared to study generalizability across radiomics implementations plus the extra features from Pyradiomics. As with the original study, this was carried out using the full intensity range in the images and thresholding to analyze the tumor intensity region only. To explore the applicability to other disease sites, a head and neck dataset was analyzed using the MATLAB toolkit and Pyradiomics. The dataset used was the HN1 dataset made publicly available in the Cancer Imaging Archive [1]. The stable features of HN1 were compared with those of RIDER.

2.1. Imaging Datasets

Two publicly available datasets were used in this study.

2.1.1. RIDER Dataset

The RIDER dataset consists of non-contrast enhanced PET/CT images from pathologically confirmed NSCLC patients scanned at the Memorial Sloan-Kettering Cancer Center, New York, United States [12]. There were 31 patients in total, and they received conventional radiotherapy. Each patient had a repeat scan 15 min after the first scan, using the same scanner and imaging protocol. These are referred to as RIDER Scan 1 and RIDER Scan 2. The image datasets were acquired using two scanner types: GE LightSpeed RT16 and GE VCT. Each CT image size was 512 by 512 pixels, with pixel sizes ranging from 0.58 mm by 0.58 mm to 0.87 mm by 0.87 mm and a slice thickness of 1.25 mm.

2.1.2. HN1 dataset

The HN1 dataset contains PET/CT images of 137 head and neck patients with squamous cell carcinoma. The patients were treated with definitive radiotherapy or concurrent chemoradiation. All patients underwent a treatment planning free-breathing 18F FDG-PET-CT scan (Biograph, SOMATOM Sensation-16 with an ECAT ACCEL PET scanner; Siemens, Erlangen, Germany), 45 min after uptake. A spiral CT (3 mm slice thickness) was performed, covering the complete thoracic region. Slice thickness: 1.5–3.0 mm; in-plane resolution: 0.9 × 0.98 mm2 to 1.09 × 1.09 mm2. The data also includes gross tumor volume (GTV) delineation by an experienced radiologist and a radiotherapy structure set. Further details are given here [1].

2.2. Texture Features Analyzed

To mimic the methodology of our previous study, the MATLAB texture analysis toolkit of Vallières et al. [13] was used to extract 43 standard features from the CT-defined GTV for the RIDER dataset. Three were first-order features, and 9 were from the gray level co-occurrence matrix (GLCM), the 13 gray level run length matrix (GLRLM), the 13 gray level size zone matrix (GLSZM) and the 5 neighboring gray tone difference matrixes (NGTDM). A full list of the features and equations defining them used for the 43 radiomics features can be found in the literature [10]. Using Pyradiomics a total of 103 features were extracted from the segmented GTV. These included: 17 first-order, 13 shape, 14 gray-level dependence matrix (GLDM), 22 GLCM, 16 GLRLM, 16 gray- GLSZM, and 5 NGTDM features.

2.3. Experimental Set-up and Statistical Analysis

The 43 features from the MATLAB toolkit were generated for both RIDER Scan 1 and Scan 2. Secondly, the results were also compared with and without an intensity threshold applied to the CT scan. The threshold used was −200 to 300 HU, as in our previous study [9]. Thirdly, stability was measured by comparing the global uniform quantizer (GUQ—with the same quantizer applied to each scan) and the individual uniform quantizer (IUQ—with the quantizer optimized for each scan). All these were uniform quantizers that quantized the intensity range of each GTV into equal width bins.
The same features were generated for the two RIDER scans using Pyradiomics with the same thresholding, and Fixed Bin Width (equivalent to Global Uniform quantizer) and Fixed Bin Count (equivalent to Individual Uniform Quantizer).
Results were compared between the MATLAB and Pyradiomics implementations and between the two RIDER scans, with and without intensity thresholds. No outcome information was available; hence, validation of the features was based on assessing the reproducibility of the rank ordering using each feature with changes to the quantization parameters for all datasets. Changes in the rank ordering indicate low reproducibility, leading to unreproducible predictions of biomarkers. In addition, the stability of the other 59 features available in Pyradiomics was also studied. Although these cannot be used to comment on the consistency of the MATLAB toolkit, it is instructive to determine if they are candidate stable features.
A feature was considered reproducible if it produced the same rank ordering for the cohort regardless of the quantization parameters. Spearman’s rank correlation, rs, was used to measure the rank ordering quantized using IUQ against GUQ at 128 intensity levels used as a reference with and without intensity thresholding, for all datasets. The rank ordering quantized with GUQ at 128 intensity levels as a reference was used as the standard comparator as it was found to be the most stable quantization combination [9]. Validation was considered successful if a feature that expressed high or low correlation in the MATLAB toolkit also expressed high or low correlation in Pyradiomics.
In the comparison of stable features between arms of the study in the results, e.g., between radiomics toolkits and disease sites, Venn diagrams are used to illustrate which features show promise as stable features in multiple arms.

3. Results

Table 1 lists all features that were reproducible, with high correlation, for RIDER Scans 1 and 2 based on the Spearman’s correlation coefficient with threshold (blue) and without threshold (red) using the two quantizers GUQ and IUQ. Features with rs ≥ 0.75 were classified as highly correlated. This was a subjective decision as no published guidance was available. A total of 34 features were found to be reproducible: 29 and 21 with and without intensity threshold applied, respectively, and features were found to be reproducible for all data. This suggests that including intensity threshold around the region of interest as a pre-processing step tends to increase the stability of some radiomics features.
Figure 1 shows a Venn diagram plotting the overlap in highly correlated features quantized with GUQ versus IUQ for 43 features that are commonly available in the MATLAB toolkit and Pyradiomics. Panels A and B plot the correlation of rank ordering quantized between GUQ and IUQ without and with the use of an intensity threshold. The correlation value for most shape, first order, GLCM, GLRLM, GLZSM, and NGTDM texture features were high (>0.8) for the RIDER Scan 1, with 9 and 14 features with rs > 0.9 without and with intensity threshold. The number of features that overlapped between the MATLAB toolkit and Pyradiomics was higher when the intensity threshold was applied. A similar trend was seen for RIDER scan 2 data.
Figure 2 presents the results for features that were uniquely available in Pyradiomics. Figure 2A shows the correlation of rank order between features extracted with and without an intensity threshold with the use of GUQ and IUQ quantization. Similarly, Figure 2B shows the correlation of ranks quantized between GUQ and IUQ with and without intensity threshold being applied. Shape features were found to be invariant to intensity thresholds or quantization techniques. For all other classes, the features showed more variance based on the application of an intensity threshold and less so on the quantization used. A few features showed negative correlation with the choice of quantization used, namely GLCM Inverse Variance, GLDM Large Dependence Low Gray Level Variance, GLRLM Run Entropy, and GLZSM Size Zone Nonuniformity Normalized. This negative correlation was seen only in the GLCM Maximum Probability when an intensity threshold was applied.
Figure 3 plots a Venn diagram with the overlap in the highly correlated features quantized between GUQ and IUQ and with an intensity threshold applied for datasets RIDER Lung 1 and Head and Neck HN1.The features were extracted using Pyradiomics; 18/43 features were agnostic of the disease site and 4/5 from HN1 and 8/11 from RIDER lung had rank correlations above 0.9, respectively.

4. Discussion

The purpose of this study was to verify the robustness of the methods reported earlier for improving the stability of radiomic features. The study used two different datasets with different pixel sizes acquired in a different center using different scanners and protocols, one for lung cancer and the other for head and neck cancer, both available publicly, while additionally testing for reproducibility and stability in the implementation of radiomic features. High levels of correlation were achieved for more than half of the features for both MATLAB and Pyradiomics implementations for the RIDER lung data, and more than one-third of the features were stable for lung and head and neck datasets, suggesting that some features are agnostic to the disease site and generally robust.
There are several characteristics of imaging systems that are relevant to their use in healthcare. These include pixel size and spatial resolution, acquisition parameters (such as kVp in x-rays and imaging sequence in MRI), tomographic reconstruction parameters, patient positioning, and when the image is taken on the patient pathway. These are all important parameters at the start of the process and are determined before radiomics analysis is performed. One of the biggest challenges in radiomics is the low reproducibility of the results from various studies [4,11]. Some of the possible causes for the low reproducibility include pre-existing differences in the dataset used, for example, different acquisition parameters [14,15], reconstruction methods [16,17], pixel sizes [18] and slice thickness; low reproducibility of features due to variations in quantization parameters; and low repeatability of the features [14]. Other considerations include the preprocessing of the imaging data [19]. For instance, Mottola et al. studied the effects of image resampling and showed that different resampling approaches produced very different error metrics, with Lanczos interpolation performing substantially better than simple linear interpolation [20]. Cui and Yin have detailed the impact that image quality has on radiomic applications and summarized the minimum image quality requirements and recommendations for reducing the impact of image quality on the reproducibility of radiomic studies [11]. Broadly, for radiomics studies, it is important that all parameter choices are documented and reported, and more specifically, it is good practice that imaging parameters are kept as consistent as possible. Binsheng Zhao suggests quantitative methods/metrics to help determine image quality and/or similarity to recognize comparable images that can be used interchangeably or to decide whether an image’s quality is adequate for computing radiomics features [21]. Often, radiomic features identified as predictive are based on small datasets, may be biased toward the specific dataset, and have limited predictive power on another dataset. For other sources of variability affecting radiomics models, readers are directed to some of these studies [19,21,22].
The aim of many radiomics studies, including those by our group, is to predict an outcome such as response to treatment or disease-free survival using one or multiple features referred to as biomarkers. Predictions are often performed using statistical approaches, including Kaplan-Meier analysis based on a single feature of the data at a time [23] and machine learning approaches with multiple features from a large set of features up to hundreds [1,24]. Outcome prediction accuracy is heavily reliant on having highly reproducible features. For instance, the widely used Kaplan-Meier analysis method involves ordering the dataset based on a feature and dichotomizing it into two sets for prediction. It is vital for the rank ordering to be consistent, as changes in the rank order may change the dichotomization and hence results in Kaplan-Meier studies, leading to low reproducibility and low predictive power. In our previous work [9], we reported a methodology to evaluate the rank order of the features and have shown that some radiomics features are reproducible across different scanner models, acquisition parameters, reconstruction methods, and modest variations in slice thickness, provided pixel sizes are resampled to a fixed standard. It was identified that feature reproducibility was highly sensitive to the choice of quantization parameters. This study has successfully validated our previous results [9] and reproduced the changes in radiomics features using different quantization parameters, suggesting the methodology used for the study is robust, even when using a different radiomics feature extraction implementation [3]. These results highlight the importance of reporting the detailed methodology used. Based on studies in the literature [11,14,25] and our own results, we recommend excluding unreproducible features from analysis to reduce dimensionality and computational burden. To improve further studies have suggested that deep learning could be considered to improve the image quality of the CT images [26], which may lead to reproducible radiomics features [27]. This will need to be explored further in future studies. In this era of deep learning, Chung et al. have raised a question for further thought for the radiology and quantitative imaging communities: have we already lost a lot of information available when we choose to reconstruct images for visual interpretation? [28].
There are some limitations to this study. Pyradiomics does not comply completely with all the recommendations of the Imaging Biomarker Standardization Initiative (IBSI) [27], for example, the quantization parameters. Although care was taken to keep the suggested stability parameters as close as possible, their implementation would have affected the present study. Hence, IBSI compliance is strongly recommended to allow better reproduction and validation of the treatment results externally [29]. The goal of all radiomics studies is to predict clinically relevant properties and/or disease outcomes, such as disease recurrence or survival. The study has only focused on the reproducibility of the features; however, the reproducibility of a feature does not automatically imply that it is clinically informative. The next stage in evaluating this methodology will be to apply it to the modeling of outcomes.

5. Conclusions

Radiomics features reported as stable were analyzed for reproducibility using the RIDER lung dataset, with 29 of 43 features found to be reproducible to changes in the feature extraction toolkits when intensity threshold was applied, maintaining stable rank ordering (rs > 0.8), and are recommended for use for biomarker analysis. We found that 18/43 reported features were common in the RIDER and HN1 datasets, suggesting they may be agnostic to disease site. Useful radiomics features should be selected based on reproducibility. This study identified a set of features that meet this requirement and validated the methodology for evaluating reproducibility between datasets.

Author Contributions

Conceptualization, H.M.T.T., E.M.D., C.P.S., A.N., V.P. and P.M.E.; methodology, H.M.T.T., H.Y.C.W., E.M.D., C.P.S., A.N. and P.M.E.; software, H.M.T.T. and H.Y.C.W.; validation, H.M.T.T., H.Y.C.W. and H.S.; formal analysis, H.M.T.T. and H.Y.C.W.; investigation, V.P., A.J.V., B.K.S., S.P.P., D.D., M.M. and R.G.I.; resources, A.N., V.P., P.M.E. and D.D.; data curation, H.M.T.T. and H.Y.C.W.; writing—original draft preparation, H.M.T.T., H.Y.C.W. and P.M.E.; writing—review and editing, H.M.T.T. and P.M.E.; visualization, V.P., A.J.V., B.K.S., S.P.P., D.D., M.M. and R.G.I.; supervision, E.M.D., C.P.S., A.N., V.P. and P.M.E.; project administration, H.M.T.T., V.P. and P.M.E.; funding acquisition, H.M.T.T., A.N., V.P. and P.M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the DBT/Wellcome Trust India Alliance Early Career Fellowship [Grant number: IA/E/18/1/504306] awarded to HMT Thomas. HMT Thomas acknowledges the support from the University of Surrey IAS fellowship for external academics in 2022-23. HYC Wang acknowledges PhD funding from Alliance Medical Ltd.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to its use of solely publicly available datasets.

Informed Consent Statement

Informed consent was not applicable for this study.

Data Availability Statement

This study used two publicly available datasets: RIDER and HN1. The RIDER dataset is available from the Cancer Imaging Archive: https://wiki.cancerimagingarchive.net/display/Public/RIDER+Lung+CT (accessed on 17 June 2023). Details of HN1 are given in reference [1]. The MATLAB code used is from the toolbox of Vallières: https://github.com/mvallieres/radiomics/tree/master/TextureToolbox (accessed on 17 June 2023). The functions used were: prepareVolume.m, equalQuantization.m, getGLCM.m, getGLSZM.m, getGLRLM.m, getNGTDM.m, and getGlobalTextures.m. The MATLAB user code is given in Appendix A. The Pyradiomics parameter file used for the feature extraction is given in Appendix B.

Conflicts of Interest

The authors have no relevant conflicts of interest to disclose.

Abbreviations

GLCMGray-Level Co-Occurrence Matrix
GLDMGray-Level Dependence Matrix
GLRLMGray Level Run Length Matrix
GLSZMGrey Level Size Zone Matrix
GTVGross Tumor Volume
GUQGlobal Uniform Quantizer
ICCInterclass Correlation
IUQIndividual Uniform Quantizer
NGTDMNeighboring Gray Tone Difference Matrix
RIDERReference Image Database to Evaluate Therapy Response
rsSpearman’s rank correlation

Appendix A. MATLAB User File

  • foldername = ‘folder with GTVseg’;
  • cd ‘to your working directory’;
  • for ns = 1:length(s)
  • filename = [‘GTVsegmentation_pt’ num2str(s(ns)) ‘.mat’];
  • files = load(fullfile(foldername,filename));
  • names=fieldnames(files);
  • for int= 1:size(names,1)
  • volume = files.(names{int,1});
  • end
  • mask=volume;
  • mask(mask>0)=1;
  •  
  • % thresholding
  • volume_low = 800 <= volume;
  • volume_high = volume <= 1300;
  • volume = volume.*volume_low;
  • volume = volume.*volume_high;
  • texture = zeros(1,43);
  •  
  • % Specify the Quantization level
  • for Qsize = 128
  • % choose quantiser
  • % https://github.com/mvallieres/radiomics/blob/master/TextureToolbox/Pre-processing/prepareVolume.m
  • GUQ
  • [ROIonly,levels] = prepareVolume(volume,mask,’Other’,’pixelW’,’sliceS’,1,’pixelW’,’Matrix’,’UniformRange’,Qsize);
  • % IUQ
  • [ROIonly,levels] = prepareVolume(volume,mask,’Other’,’pixelW’,’sliceS’,1,’pixelW’,’Matrix’,’Uniform’,Qsize);
  •  
  • ROIOnly = ROIonly;
  •  
  • % generate texture features using toolbox
  • % https://github.com/mvallieres/radiomics/tree/master/TextureToolbox/GLCM/getGLCM.m
  • % implements quanitiser from toolbox
  • % https://github.com/mvallieres/radiomics/blob/master/TextureToolbox/Pre-processing/Quantization/equalQuantization.m
  • [GLCM] = getGLCM(ROIonly,levels);
  • [textures] = getGLCMtextures(GLCM);
  • texture(Qsize,1:9) = [textures.Energy textures.Contrast textures.Entropy textures.Homogeneity textures.Correlation textures.SumAverage textures.Variance textures.Dissimilarity textures.AutoCorrelation];
  •  
  • % https://github.com/mvallieres/radiomics/tree/master/TextureToolbox/GLSZM/getGLSZM.m
  • [GLSZM] = getGLSZM(ROIOnly,levels);
  • [textures] = getGLSZMtextures(GLSZM);
  • texture(Qsize,10:22) = [textures.SZE textures.LZE textures.GLN textures.ZSN textures.ZP textures.LGZE textures.HGZE textures.SZLGE textures.SZHGE textures.LZLGE textures.LZHGE textures.GLV textures.ZSV];
  •  
  • % https://github.com/mvallieres/radiomics/tree/master/TextureToolbox/GLRLM/getGLRLM.m
  • [GLRLM] = getGLRLM(ROIonly,levels);
  • [textures] = getGLRLMtextures(GLRLM);
  • texture(Qsize,23:35) = [textures.SRE textures.LRE textures.GLN textures.RLN textures.RP textures.LGRE textures.HGRE textures.SRLGE textures.SRHGE textures.LRLGE textures.LRHGE textures.GLV textures.RLV];
  •  
  • % https://github.com/mvallieres/radiomics/tree/master/TextureToolbox/NGTDM/getNGTDM.m
  • [NGTDM,countValid] = getNGTDM(ROIOnly,levels);
  • [textures] = getNGTDMtextures(NGTDM,countValid);
  • texture(Qsize,36:40) = [textures.Coarseness textures.Contrast textures.Busyness textures.Complexity textures.Strength];
  •  
  • % https://github.com/mvallieres/radiomics/tree/master/TextureToolbox/Global/getGlobalTextures.m
  • [textures] = getGlobalTextures(ROIonly,Qsize);
  • texture(Qsize,41:43) = [textures.Variance textures.Skewness textures.Kurtosis];
  •  
  • % save after each quantisation level
  • save(([‘texture_pt’ num2str(s(ns)) ‘.mat’]),’-v7.3’,’texture’);
  • disp([‘pt’ num2str(s(ns)) ‘ n’ num2str(Qsize)])
  •  
  • end

Appendix B. Pyradiomics Parameter File

  • Extracted using PyRadiomics version: 2.1.0
  • imageType:
  • Original:
  • binCount: 128
  • featureClass:
  • glcm:
  • - ‘JointEnergy’
  • - ‘Contrast’
  • - ‘JointEntropy’
  • - ‘Id’
  • - ‘Correlation’
  • - ‘SumAverage’
  • - ‘SumSquares’
  • - ‘DifferenceAverage’
  • - ‘Autocorrelation’
  • glszm:
  • - ‘SmallAreaEmphasis’
  • - ‘LargeAreaEmphasis’
  • - ‘GrayLevelNonUniformity’
  • - ‘SizeZoneNonUniformity’
  • - ‘ZonePercentage’
  • - ‘LowGrayLevelZoneEmphasis’
  • - ‘HighGrayLevelZoneEmphasis’
  • - ‘SmallAreaLowGrayLevelEmphasis’
  • - ‘SmallAreaHighGrayLevelEmphasis’
  • - ‘LargeAreaLowGrayLevelEmphasis’
  • - ‘LargeAreaHighGrayLevelEmphasis’
  • - ‘GrayLevelVariance’
  • - ‘ZoneVariance’
  • glrlm:
  • - ‘ShortRunEmphasis’
  • - ‘LongRunEmphasis’
  • - ‘GrayLevelNonUniformity’
  • - ‘RunLengthNonUniformity’
  • - ‘RunPercentage’
  • - ‘LowGrayLevelRunEmphasis’
  • - ‘HighGrayLevelRunEmphasis’
  • - ‘ShortRunLowGrayLevelEmphasis’
  • - ‘ShortRunHighGrayLevelEmphasis’
  • - ‘LongRunLowGrayLevelEmphasis’
  • - ‘LongRunHighGrayLevelEmphasis’
  • - ‘GrayLevelVariance’
  • - ‘RunVariance’
  •  
  • ngtdm:
  • - ‘Coarseness’
  • - ‘Contrast’
  • - ‘Busyness’
  • - ‘Complexity’
  • - ‘Strength’
  •  
  • firstorder: # Remove Total Energy, correlated to Energy (due to resampling enabled)
  • - ‘Variance’
  • - ‘Skewness’
  • - ‘Kurtosis’
  •  
  • setting:
  • # Resampling:
  • interpolator: ‘sitkLinear’
  • resampledPixelSpacing: [0.98, 0.98, 0]
  • resegmentRange: [−200, 300]
  • resegmentMode: absolute
  • # Misc:
  • label: 1

References

  1. Aerts, H.J.W.L.; Velazquez, E.R.; Leijenaar, R.T.H.; Parmar, C.; Grossmann, P.; Carvalho, S.; Bussink, J.; Monshouwer, R.; Haibe-Kains, B.; Rietveld, D.; et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 2014, 5, 1–9. [Google Scholar] [CrossRef] [Green Version]
  2. O’Sullivan, B.; Brierley, J.; Byrd, D.; Bosman, F.; Kehoe, S.; Kossary, C.; Piñeros, M.; Van Eycken, E.; Weir, H.K.; Gospodarowicz, M. The TNM classification of malignant tumours—Towards common understanding and reasonable expectations. Lancet Oncol. 2017, 18, 849–851. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chalmers, I.; Glasziou, P. Avoidable waste in the production and reporting of research evidence. Lancet 2009, 374, 86–89. [Google Scholar] [CrossRef] [PubMed]
  4. Traverso, A.; Wee, L.; Dekker, A.; Gillies, R. Repeatability and Reproducibility of Radiomic Features: A Systematic Review. Int. J. Radiat. Oncol. Biol. Phys. 2018, 102, 1143–1158. [Google Scholar] [CrossRef] [Green Version]
  5. Balagurunathan, Y.; Gu, Y.; Wang, H.; Kumar, V.; Grove, O.; Hawkins, S.; Kim, J.; Goldgof, D.B.; Hall, L.O.; Gatenby, R.A.; et al. Reproducibility and Prognosis of Quantitative Features Extracted from CT Images. Transl. Oncol. 2014, 7, 72–87. [Google Scholar] [CrossRef] [Green Version]
  6. Coroller, T.P.; Agrawal, V.; Narayan, V.; Hou, Y.; Grossmann, P.; Lee, S.W.; Mak, R.H.; Aerts, H.J.W.L. Radiomic phenotype features predict pathological response in non-small cell lung cancer. Radiother. Oncol. 2016, 119, 480–486. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Fave, X.; Mackin, D.; Yang, J.; Zhang, J.; Fried, D.; Balter, P.; Followill, D.; Gomez, D.; Jones, A.K.; Stingo, F.; et al. Can radiomics features be reproducibly measured from CBCT images for patients with non-small cell lung cancer? Med. Phys. 2015, 42, 6784–6797. [Google Scholar] [CrossRef] [Green Version]
  8. Fave, X.; Zhang, L.; Yang, J.; Mackin, D.; Balter, P.; Gomez, D.; Followill, D.; Jones, A.K.; Stingo, F.; Liao, Z.; et al. Delta-radiomics features for the prediction of patient outcomes in non–small cell lung cancer. Sci. Rep. 2017, 7, 588. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, H.Y.C.; Donovan, E.M.; Nisbet, A.; South, C.P.; Alobaidli, S.; Ezhil, V.; Phillips, I.; Prakash, V.; Ferreira, M.; Webster, P.; et al. The stability of imaging biomarkers in radiomics: A framework for evaluation. Phys. Med. Biol. 2019, 64, 165012. [Google Scholar] [CrossRef]
  10. Vallières, M.; Kay-Rivest, E.; Perrin, L.J.; Liem, X.; Furstoss, C.; Aerts, H.J.W.L.; Khaouam, N.; Nguyen-Tan, P.F.; Wang, C.-S.; Sultanem, K.; et al. Radiomics strategies for risk assessment of tumour failure in head-and-neck cancer. Sci. Rep. 2017, 7, 10117. [Google Scholar] [CrossRef] [Green Version]
  11. Cui, Y.; Yin, F.-F. Impact of image quality on radiomics applications. Phys. Med. Biol. 2022, 67, 15TR03. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, B.; James, L.P.; Moskowitz, C.S.; Guo, P.; Ginsberg, M.S.; Lefkowitz, R.A.; Qin, Y.; Riely, G.J.; Kris, M.G.; Schwartz, L.H.; et al. Evaluating variability in tumor measurements from same-day repeat CT scans of patients with non-small cell lung cancer. Radiology 2009, 252, 263–272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Welcome to Pyradiomics Documentation!—Pyradiomics v3.0.1.post15+g2791e23 Documentation n.d. Available online: https://pyradiomics.readthedocs.io/en/latest/ (accessed on 27 January 2023).
  14. Jha, A.K.; Mithun, S.; Jaiswar, V.; Sherkhane, U.B.; Purandare, N.C.; Prabhash, K.; Rangarajan, V.; Dekker, A.; Wee, L.; Traverso, A. Repeatability and reproducibility study of radiomic features on a phantom and human cohort. Sci. Rep. 2021, 11, 2055. [Google Scholar] [CrossRef] [PubMed]
  15. Buch, K.; Li, B.; Qureshi, M.M.; Kuno, H.; Anderson, S.W.; Sakai, O. Quantitative Assessment of Variation in CT Parameters on Texture Features: Pilot Study Using a Nonanatomic Phantom. Am. J. Neuroradiol. 2017, 38, 981–985. [Google Scholar] [CrossRef] [Green Version]
  16. Berenguer, R.; del Rosario Pastor-Juan, M.; Canales-Vázquez, J.; Castro-García, M.; Villas, M.V.; Mansilla Legorburo, F.; Sabater, S. Radiomics of CT Features May Be Nonreproducible and Redundant: Influence of CT Acquisition Parameters. Radiology 2018, 288, 407–415. [Google Scholar] [CrossRef] [Green Version]
  17. Ahn, S.J.; Kim, J.H.; Lee, S.M.; Park, S.J.; Han, J.K. CT reconstruction algorithms affect histogram and texture analysis: Evidence for liver parenchyma, focal solid liver lesions, and renal cysts. Eur. Radiol. 2019, 29, 4008–4015. [Google Scholar] [CrossRef] [PubMed]
  18. Shafiq-ul-Hassan, M.; Latifi, K.; Zhang, G.; Ullah, G.; Gillies, R.; Moros, E. Voxel size and gray level normalization of CT radiomic features in lung cancer. Sci. Rep. 2018, 8, 10545. [Google Scholar] [CrossRef] [Green Version]
  19. Varghese, A.J.; Gouthamchand, V.; Sasidharan, B.K.; Wee, L.; Sidhique, S.K.; Rao, J.P.; Dekker, A.; Hoebers, F.; Devakumar, D.; Irodi, A.; et al. Multi-centre radiomics for prediction of recurrence following radical radiotherapy for head and neck cancers: Consequences of feature selection, machine learning classifiers and batch-effect harmonization. Phys. Imaging Radiat. Oncol. 2023, 26, 100450. [Google Scholar] [CrossRef]
  20. Mottola, M.; Ursprung, S.; Rundo, L.; Sanchez, L.E.; Klatte, T.; Mendichovszky, I.; Stewart, G.D.; Sala, E.; Bevilacqua, A. Reproducibility of CT-based radiomic features against image resampling and perturbations for tumour and healthy kidney in renal cancer patients. Sci. Rep. 2021, 11, 11542. [Google Scholar] [CrossRef] [PubMed]
  21. Zhao, B. Understanding Sources of Variation to Improve the Reproducibility of Radiomics. Front. Oncol. 2021, 11, 826. [Google Scholar] [CrossRef]
  22. van Timmeren, J.E.; Cester, D.; Tanadini-Lang, S.; Alkadhi, H.; Baessler, B. Radiomics in medical imaging—“how-to” guide and critical reflection. Insights Into Imaging 2020, 11, 91. [Google Scholar] [CrossRef] [PubMed]
  23. Win, T.; Miles, K.A.; Janes, S.M.; Ganeshan, B.; Shastry, M.; Endozo, R.; Meagher, M.; Shortman, R.I.; Wan, S.; Kayani, I.; et al. Tumor heterogeneity and permeability as measured on the CT component of PET/CT predict survival in patients with non-small cell lung cancer. Clin. Cancer Res. 2013, 19, 3591–3599. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Coroller, T.P.; Agrawal, V.; Huynh, E.; Narayan, V.; Lee, S.W.; Mak, R.H.; Aerts, H.J.W.L. Radiomic-Based Pathological Response Prediction from Primary Tumors and Lymph Nodes in NSCLC. J. Thorac. Oncol. 2017, 12, 467–476. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Adeoye, J.; Hui, L.; Su, Y.-X. Data-centric artificial intelligence in oncology: A systematic review assessing data quality in machine learning models for head and neck cancer. J. Big Data 2023, 10, 28. [Google Scholar] [CrossRef]
  26. Li, D.; Ma, L.; Li, J.; Qi, S.; Yao, Y.; Teng, Y. A comprehensive survey on deep learning techniques in CT image quality improvement. Med. Biol. Eng. Comput. 2022, 60, 2757–2770. [Google Scholar] [CrossRef]
  27. Yang, B.; Chen, X.; Yuan, S.; Liu, Y.; Dai, J.; Men, K. Deep learning improves image quality and radiomics reproducibility for high-speed four-dimensional computed tomography reconstruction. Radiother. Oncol. 2022, 170, 184–189. [Google Scholar] [CrossRef]
  28. Chung, C.; Kalpathy-Cramer, J.; Knopp, M.V.; Jaffray, D.A. In the Era of Deep Learning, Why Reconstruct an Image at All? J. Am. Coll. Radiol. 2021, 18, 170–173. [Google Scholar] [CrossRef]
  29. Hatt, M.; Vallieres, M.; Visvikis, D.; Zwanenburg, A. IBSI: An international community radiomics standardization initiative. J. Nucl. Med. 2018, 59, 287. [Google Scholar]
Figure 1. Venn diagrams with a visual representation of features that are reproducible between MATLAB and Pyradiomics feature extraction toolkits for GUQ versus IUQ with 128 quantizer levels (Fixed Bin Width in Pyradiomics). (A) without threshold; (B) with threshold applied. Reproducibility is measured by an rs value greater than or equal to 0.8. Features highlighted in bold had rs value greater than 0.9.
Figure 1. Venn diagrams with a visual representation of features that are reproducible between MATLAB and Pyradiomics feature extraction toolkits for GUQ versus IUQ with 128 quantizer levels (Fixed Bin Width in Pyradiomics). (A) without threshold; (B) with threshold applied. Reproducibility is measured by an rs value greater than or equal to 0.8. Features highlighted in bold had rs value greater than 0.9.
Applsci 13 07291 g001
Figure 2. The Spearman correlation ranks of features extracted using (A) threshold versus no threshold and comparing the impact of the quantization techniques (GUQ (bin count) vs. IUQ (bin width)) and (B) correlation between quantization methods (GUQ and IUQ) and comparing the impact of thresholding. Each color-paired block represents a separate feature class.
Figure 2. The Spearman correlation ranks of features extracted using (A) threshold versus no threshold and comparing the impact of the quantization techniques (GUQ (bin count) vs. IUQ (bin width)) and (B) correlation between quantization methods (GUQ and IUQ) and comparing the impact of thresholding. Each color-paired block represents a separate feature class.
Applsci 13 07291 g002
Figure 3. Venn diagram with a visual representation of features that are reproducible between RIDER Lung 1 and the Head and Neck (HN1) dataset and with an intensity threshold applied. Reproducibility is measured by an rs value greater than or equal to 0.8. Features highlighted in bold had an rs value greater than 0.9.
Figure 3. Venn diagram with a visual representation of features that are reproducible between RIDER Lung 1 and the Head and Neck (HN1) dataset and with an intensity threshold applied. Reproducibility is measured by an rs value greater than or equal to 0.8. Features highlighted in bold had an rs value greater than 0.9.
Applsci 13 07291 g003
Table 1. List of radiomics features with reproducibility with threshold and without threshold shown in blue and black for RIDER Scans 1 and 2, respectively, in columns 2, 3, 4 and 5 for MATLAB, columns 6, 7, 8, and 9 for Pyradiomics. Spearman correlation coefficient values ≤0.75 is shown in red. Columns 10 and 11 represent the reproducible features across both scans and implementations; with threshold, are shown in orange tick and without threshold in purple tick, respectively; across all data is shown by a green tick in column 12. Colours are explained in the text.
Table 1. List of radiomics features with reproducibility with threshold and without threshold shown in blue and black for RIDER Scans 1 and 2, respectively, in columns 2, 3, 4 and 5 for MATLAB, columns 6, 7, 8, and 9 for Pyradiomics. Spearman correlation coefficient values ≤0.75 is shown in red. Columns 10 and 11 represent the reproducible features across both scans and implementations; with threshold, are shown in orange tick and without threshold in purple tick, respectively; across all data is shown by a green tick in column 12. Colours are explained in the text.
Radiomics FeatureMatlabPyradiomicsAcross Scans 1 & 2
and Feature Extraction
Implementations
All Data
Scan 1Scan 2Scan 1Scan 2Scan 1Scan 2Scan 1Scan 2ThresholdW/o
Threshold
ThresholdW/o ThresholdThresholdW/o Threshold
Variance1.001.001.000.991.001.001.001.00
Skewness1.001.001.000.991.001.001.001.00
Kurtosis1.001.001.000.981.001.001.001.00
GLCM Energy0.950.960.730.830.160.000.710.00
GLCM Contrast1.001.000.800.810.850.950.910.95
GLCM Entropy0.980.970.790.84−0.03−0.100.50−0.10
GLCM Homogeneity0.960.980.690.750.930.970.940.97
GLCM Correlation1.001.001.000.980.920.981.000.98
GLCM Sum Average1.001.000.180.270.950.900.080.90
GLCM Variance1.001.000.800.810.870.980.920.98
GLCM Autocorrelation1.001.000.150.230.940.880.060.88
GLCM Dissimilarity1.001.000.800.820.890.970.940.97
GLSZM SZE0.800.770.830.84−0.42−0.17−0.54−0.17
GLSZM LZE0.940.900.750.820.760.740.880.74
GLSZM GLN0.980.980.880.860.980.990.970.99
GLSZM ZSN0.810.770.830.840.990.980.960.98
GLSZM ZP0.870.880.780.860.790.790.790.79
GLSZM LGZE0.990.990.840.810.390.230.790.23
GLSZM HGZE0.970.980.170.210.220.34−0.070.34
GLSZM SZLGE0.990.990.870.81−0.060.140.780.14
GLSZM SZHGE0.740.790.060.050.240.35−0.210.35
GLSZM LZLGE0.991.000.550.43−0.72−0.700.34−0.70
GLSZM LZHGE0.920.910.930.840.490.570.870.57
GLSZM GLV0.980.990.650.710.01−0.12−0.22−0.12
GLSZM ZSV0.980.930.810.740.730.730.870.73
GLRLM SRE0.970.970.750.800.760.950.940.95
GLRLM LRE0.970.990.750.790.970.960.900.96
GLRLM GLN0.930.960.820.900.990.990.990.99
GLRLM RLN0.970.970.750.800.990.990.990.99
GLRLM RP0.970.980.750.800.980.970.920.97
GLRLM LGRE1.001.000.870.860.860.880.880.88
GLRLM HGRE1.001.000.130.190.960.830.010.83
GLRLM SRLGE1.001.000.880.870.930.950.860.95
GLRLM SRHGE0.991.000.150.180.830.69−0.030.69
GLRLM LRLGE1.001.000.880.85−0.41−0.360.45−0.36
GLRLM LRHGE0.990.990.210.390.180.300.330.30
GLRLM GLV1.000.990.630.760.760.870.850.87
GLRLM RLV0.940.950.820.730.960.950.910.95
NGTDM Coarseness0.990.991.000.961.001.001.001.00
NGTDM Contrast1.001.000.960.960.970.970.980.97
NGTDM Busyness0.990.990.940.931.000.990.970.99
NGTDM Complexity1.001.00−0.30−0.200.710.73−0.100.73
NGTDM Strength1.001.000.930.921.001.000.961.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Thomas, H.M.T.; Wang, H.Y.C.; Varghese, A.J.; Donovan, E.M.; South, C.P.; Saxby, H.; Nisbet, A.; Prakash, V.; Sasidharan, B.K.; Pavamani, S.P.; et al. Reproducibility in Radiomics: A Comparison of Feature Extraction Methods and Two Independent Datasets. Appl. Sci. 2023, 13, 7291. https://doi.org/10.3390/app13127291

AMA Style

Thomas HMT, Wang HYC, Varghese AJ, Donovan EM, South CP, Saxby H, Nisbet A, Prakash V, Sasidharan BK, Pavamani SP, et al. Reproducibility in Radiomics: A Comparison of Feature Extraction Methods and Two Independent Datasets. Applied Sciences. 2023; 13(12):7291. https://doi.org/10.3390/app13127291

Chicago/Turabian Style

Thomas, Hannah Mary T., Helen Y. C. Wang, Amal Joseph Varghese, Ellen M. Donovan, Chris P. South, Helen Saxby, Andrew Nisbet, Vineet Prakash, Balu Krishna Sasidharan, Simon Pradeep Pavamani, and et al. 2023. "Reproducibility in Radiomics: A Comparison of Feature Extraction Methods and Two Independent Datasets" Applied Sciences 13, no. 12: 7291. https://doi.org/10.3390/app13127291

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop