TY - JOUR
T1 - Impact of interobserver variability in manual segmentation of non‐small cell lung cancer (Nsclc) applying low‐rank radiomic representation on computed tomography
AU - Hershman, Michelle
AU - Yousefi, Bardia
AU - Serletti, Lacey
AU - Galperin‐aizenberg, Maya
AU - Roshkovan, Leonid
AU - Luna, José Marcio
AU - Thompson, Jeffrey C.
AU - Aggarwal, Charu
AU - Carpenter, Erica L.
AU - Kontos, Despina
AU - Katz, Sharyn I.
N1 - Publisher Copyright:
© 2021 by the authors. Licensee MDPI, Basel, Switzerland.
PY - 2021/12/1
Y1 - 2021/12/1
N2 - This study tackles interobserver variability with respect to specialty training in manual segmentation of non‐small cell lung cancer (NSCLC). Four readers included for segmentation are: a data scientist (BY), a medical student (LS), a radiology trainee (MH), and a specialty‐trained radiologist (SK) for a total of 295 patients from two publicly available databases. Sørensen–Dice (SD) coefficients and low rank Pearson correlation coefficients (CC) of 429 radiomics were calculated to assess interobserver variability. Cox proportional hazard (CPH) models and Kaplan‐Meier (KM) curves of overall survival (OS) prediction for each dataset were also generated. SD and CC for segmentations demonstrated high similarities, yielding, SD: 0.79 and CC: 0.92 (BY‐SK), SD: 0.81 and CC: 0.83 (LS‐SK), and SD: 0.84 and CC: 0.91 (MH‐SK) in average for both databases, respectively. OS through the maximal CPH model for the two datasets yielded c‐statistics of 0.7 (95% CI) and 0.69 (95% CI), while adding radiomic and clinical variables (sex, stage/morphological status, and histology) together. KM curves also showed significant discrimination between high‐ and low‐risk patients (p‐value < 0.005). This supports that readers’ level of training and clinical experience may not significantly influence the ability to extract accurate radiomic features for NSCLC on CT. This potentially allows flexibility in the training required to produce robust prognostic imaging biomarkers for potential clinical translation.
AB - This study tackles interobserver variability with respect to specialty training in manual segmentation of non‐small cell lung cancer (NSCLC). Four readers included for segmentation are: a data scientist (BY), a medical student (LS), a radiology trainee (MH), and a specialty‐trained radiologist (SK) for a total of 295 patients from two publicly available databases. Sørensen–Dice (SD) coefficients and low rank Pearson correlation coefficients (CC) of 429 radiomics were calculated to assess interobserver variability. Cox proportional hazard (CPH) models and Kaplan‐Meier (KM) curves of overall survival (OS) prediction for each dataset were also generated. SD and CC for segmentations demonstrated high similarities, yielding, SD: 0.79 and CC: 0.92 (BY‐SK), SD: 0.81 and CC: 0.83 (LS‐SK), and SD: 0.84 and CC: 0.91 (MH‐SK) in average for both databases, respectively. OS through the maximal CPH model for the two datasets yielded c‐statistics of 0.7 (95% CI) and 0.69 (95% CI), while adding radiomic and clinical variables (sex, stage/morphological status, and histology) together. KM curves also showed significant discrimination between high‐ and low‐risk patients (p‐value < 0.005). This supports that readers’ level of training and clinical experience may not significantly influence the ability to extract accurate radiomic features for NSCLC on CT. This potentially allows flexibility in the training required to produce robust prognostic imaging biomarkers for potential clinical translation.
KW - Computed tomography (CT)
KW - Interobserver variability
KW - Non‐small cell lung cancer
KW - Radiomics
UR - http://www.scopus.com/inward/record.url?scp=85120005053&partnerID=8YFLogxK
U2 - 10.3390/cancers13235985
DO - 10.3390/cancers13235985
M3 - Article
C2 - 34885094
AN - SCOPUS:85120005053
SN - 2072-6694
VL - 13
JO - Cancers
JF - Cancers
IS - 23
M1 - 5985
ER -