TY - JOUR
T1 - Weaving attention U-net
T2 - A novel hybrid CNN and attention-based method for organs-at-risk segmentation in head and neck CT images
AU - Zhang, Zhuangzhuang
AU - Zhao, Tianyu
AU - Gay, Hiram
AU - Zhang, Weixiong
AU - Sun, Baozhou
N1 - Funding Information:
We would like to thank Varian Medical System for financial support through a research grant.
Publisher Copyright:
© 2021 American Association of Physicists in Medicine
PY - 2021/11
Y1 - 2021/11
N2 - Purpose: In radiotherapy planning, manual contouring is labor-intensive and time-consuming. Accurate and robust automated segmentation models improve the efficiency and treatment outcome. We aim to develop a novel hybrid deep learning approach, combining convolutional neural networks (CNNs) and the self-attention mechanism, for rapid and accurate multi-organ segmentation on head and neck computed tomography (CT) images. Methods: Head and neck CT images with manual contours of 115 patients were retrospectively collected and used. We set the training/validation/testing ratio to 81/9/25 and used the 10-fold cross-validation strategy to select the best model parameters. The proposed hybrid model segmented 10 organs-at-risk (OARs) altogether for each case. The performance of the model was evaluated by three metrics, that is, the Dice Similarity Coefficient (DSC), Hausdorff distance 95% (HD95), and mean surface distance (MSD). We also tested the performance of the model on the head and neck 2015 challenge dataset and compared it against several state-of-the-art automated segmentation algorithms. Results: The proposed method generated contours that closely resemble the ground truth for 10 OARs. On the head and neck 2015 challenge dataset, the DSC scores of these OARs were 0.91 (Formula presented.) 0.02, 0.73 (Formula presented.) 0.10, 0.95 (Formula presented.) 0.03, 0.76 (Formula presented.) 0.08, 0.79 (Formula presented.) 0.05, 0.87 (Formula presented.) 0.05, 0.86 (Formula presented.) 0.08, 0.87 (Formula presented.) 0.03, and 0.87 (Formula presented.) 0.07 for brain stem, chiasm, mandible, left/right optic nerve, left/right submandibular, and left/right parotid, respectively. Our results of the new weaving attention U-net (WAU-net) demonstrate superior or similar performance on the segmentation of head and neck CT images. Conclusions: We developed a deep learning approach that integrates the merits of CNNs and the self-attention mechanism. The proposed WAU-net can efficiently capture local and global dependencies and achieves state-of-the-art performance on the head and neck multi-organ segmentation task.
AB - Purpose: In radiotherapy planning, manual contouring is labor-intensive and time-consuming. Accurate and robust automated segmentation models improve the efficiency and treatment outcome. We aim to develop a novel hybrid deep learning approach, combining convolutional neural networks (CNNs) and the self-attention mechanism, for rapid and accurate multi-organ segmentation on head and neck computed tomography (CT) images. Methods: Head and neck CT images with manual contours of 115 patients were retrospectively collected and used. We set the training/validation/testing ratio to 81/9/25 and used the 10-fold cross-validation strategy to select the best model parameters. The proposed hybrid model segmented 10 organs-at-risk (OARs) altogether for each case. The performance of the model was evaluated by three metrics, that is, the Dice Similarity Coefficient (DSC), Hausdorff distance 95% (HD95), and mean surface distance (MSD). We also tested the performance of the model on the head and neck 2015 challenge dataset and compared it against several state-of-the-art automated segmentation algorithms. Results: The proposed method generated contours that closely resemble the ground truth for 10 OARs. On the head and neck 2015 challenge dataset, the DSC scores of these OARs were 0.91 (Formula presented.) 0.02, 0.73 (Formula presented.) 0.10, 0.95 (Formula presented.) 0.03, 0.76 (Formula presented.) 0.08, 0.79 (Formula presented.) 0.05, 0.87 (Formula presented.) 0.05, 0.86 (Formula presented.) 0.08, 0.87 (Formula presented.) 0.03, and 0.87 (Formula presented.) 0.07 for brain stem, chiasm, mandible, left/right optic nerve, left/right submandibular, and left/right parotid, respectively. Our results of the new weaving attention U-net (WAU-net) demonstrate superior or similar performance on the segmentation of head and neck CT images. Conclusions: We developed a deep learning approach that integrates the merits of CNNs and the self-attention mechanism. The proposed WAU-net can efficiently capture local and global dependencies and achieves state-of-the-art performance on the head and neck multi-organ segmentation task.
KW - attention mechanism
KW - convolutional neural networks
KW - deep learning
KW - head and neck radiotherapy
KW - multi-organ segmentation
UR - http://www.scopus.com/inward/record.url?scp=85118172614&partnerID=8YFLogxK
U2 - 10.1002/mp.15287
DO - 10.1002/mp.15287
M3 - Article
C2 - 34655077
AN - SCOPUS:85118172614
SN - 0094-2405
VL - 48
SP - 7052
EP - 7062
JO - Medical physics
JF - Medical physics
IS - 11
ER -