TY - GEN
T1 - Breast density assessment via deep learning
T2 - Medical Imaging 2024: Computer-Aided Diagnosis
AU - Anant, Krisha
AU - Lopez, Juanita Hernandez
AU - Gupta, Sneha Das
AU - Bennett, Debbie L.
AU - Gastounioti, Aimilia
N1 - Publisher Copyright:
© 2024 SPIE.
PY - 2024
Y1 - 2024
N2 - To enhance reproducibility and robustness in mammographic density assessment, various deep learning models have been proposed to automatically classify mammographic images into BI-RADS density categories. Our study aims to compare the performances of different deep learning models in making breast density classifications from full-field digital mammography (FFDM) versus synthetic mammography (SM), the newer 2D mammographic image format acquired with digital breast tomosynthesis (DBT). We retrospectively analyzed negative (BI-RADS 1 or 2) routine mammographic screening exams (Selenia or Selenia Dimensions; Hologic) acquired at sites within the Barnes-Jewish/Christian (BJC) Healthcare network in St. Louis, MO from 2015 to 2018. BI-RADS breast density assessments of radiologists were obtained from BJC’s mammography reporting software (Magview 7.1). For each mammographic imaging modality, a balanced dataset of 4,000 women was selected so there were equal numbers of women in each of the four BI-RADS density categories, and each woman had at least one mediolateral oblique (MLO) and one craniocaudal (CC) view per breast in that mammographic imaging modality. Previously validated pre-processing steps were applied to all FFDM and SM images to standardize image orientation and intensity. Images were then split into training, validation, and test sets at ratios of 80%, 10%, and 10%, respectively, while maintaining the distribution of breast density categories and ensuring that all images of the same woman appear only in one set. ResNet-50 and EfficientNet-B0 architectures were optimized, trained, and evaluated separately for different imaging modalities. Overall, the models had comparable performance, though ResNet-50 performed slightly better in most cases. Furthermore, FFDM images had better classification accuracies than SM images. Our preliminary findings suggest that further deep learning developments and optimizations may be needed as we develop breast density deep learning models for the newer mammographic imaging modality, DBT.
AB - To enhance reproducibility and robustness in mammographic density assessment, various deep learning models have been proposed to automatically classify mammographic images into BI-RADS density categories. Our study aims to compare the performances of different deep learning models in making breast density classifications from full-field digital mammography (FFDM) versus synthetic mammography (SM), the newer 2D mammographic image format acquired with digital breast tomosynthesis (DBT). We retrospectively analyzed negative (BI-RADS 1 or 2) routine mammographic screening exams (Selenia or Selenia Dimensions; Hologic) acquired at sites within the Barnes-Jewish/Christian (BJC) Healthcare network in St. Louis, MO from 2015 to 2018. BI-RADS breast density assessments of radiologists were obtained from BJC’s mammography reporting software (Magview 7.1). For each mammographic imaging modality, a balanced dataset of 4,000 women was selected so there were equal numbers of women in each of the four BI-RADS density categories, and each woman had at least one mediolateral oblique (MLO) and one craniocaudal (CC) view per breast in that mammographic imaging modality. Previously validated pre-processing steps were applied to all FFDM and SM images to standardize image orientation and intensity. Images were then split into training, validation, and test sets at ratios of 80%, 10%, and 10%, respectively, while maintaining the distribution of breast density categories and ensuring that all images of the same woman appear only in one set. ResNet-50 and EfficientNet-B0 architectures were optimized, trained, and evaluated separately for different imaging modalities. Overall, the models had comparable performance, though ResNet-50 performed slightly better in most cases. Furthermore, FFDM images had better classification accuracies than SM images. Our preliminary findings suggest that further deep learning developments and optimizations may be needed as we develop breast density deep learning models for the newer mammographic imaging modality, DBT.
KW - artificial intelligence
KW - deep learning
KW - digital mammogram
KW - mammographic density
KW - synthetic mammogram
UR - http://www.scopus.com/inward/record.url?scp=85191513068&partnerID=8YFLogxK
U2 - 10.1117/12.3008648
DO - 10.1117/12.3008648
M3 - Conference contribution
AN - SCOPUS:85191513068
T3 - Progress in Biomedical Optics and Imaging - Proceedings of SPIE
BT - Medical Imaging 2024
A2 - Chen, Weijie
A2 - Astley, Susan M.
PB - SPIE
Y2 - 19 February 2024 through 22 February 2024
ER -