Impact Factor:4.506
DOI number:10.1002/mp.15852
Journal:Medical Physics
Abstract:Purpose: Breast cancer is the most commonly occurring cancer worldwide. The ultrasound reflectivity imaging technique can be used to obtain breast ultrasound (BUS) images, which can be used to classify benign and malignant tumors. However, the classification is subjective and dependent on the experience and skill of operators and doctors. The automatic classification method can assist doctors and improve the objectivity, but current convolution neural network (CNN) is not good at learning global features and vision transform (ViT) is not good at extraction local features. In this study, we proposed an VGG attention vision transformer (VGGA-ViT) network to overcome their disadvantages. Methods: In the proposed method, we used a convolutional neural network (CNN) module to extract the local features and employed a vision transformer (ViT) module to learn the global relationship between different regions and enhance the relevant local features. The CNN module was named the VGG attention (VGGA) module. It was composed of a visual geometry group (VGG) backbone, a feature extraction fully connected layer, and a squeeze-and-excitation (SE) block. Both the VGG backbone and the ViT module were pre-trained on the ImageNet dataset and re-trained using BUS samples in this study. Two BUS datasets were employed for validation. Results: Cross-validation was conducted on two BUS datasets. For the Dataset A, the proposed VGGA-ViT network achieved high accuracy (88.71±1.55%), recall (90.73±1.57%), specificity (85.58±3.35%), precision (90.77±1.98%), F1 score (90.73±1.24%), and Matthews correlation coefficient (MCC) (76.34±3.29%), which were better than those of all compared previous networks in this study. The Dataset B was used as a separate test set, the test results showed that the VGGA-ViT had highest accuracy (81.72±2.99%), recall (64.45±2.96%), specificity (90.28± 3.51%), precision (77.08±7.21%), F1 score (70.11±4.25%), and MCC (57.64±6.88%). Conclusions: In this study, we proposed the VGGA-ViT for the BUS classification, which was good at learning both local and global features. The proposed network achieved higher accuracy than the compared previous methods.
Translation or Not:no
Date of Publication:2022-06-22
Included Journals:SCI
Attachments:
Supervisor of Doctorate Candidates
Supervisor of Master's Candidates
E-Mail:
Date of Employment:2017-05-01
School/Department:School of Instrumentation and Optoelectronic Engineering
Administrative Position:Vice Dean of Department
Business Address:New building B504, School of Instrumentation and Optoelectronic Engineering, Beihang University
Gender:Male
Contact Information:quxiaolei@gmail.com
Status:Employed
Academic Titles:Associate professor
Alma Mater:the University of Tokyo
Discipline:Instrumentation Science and Technology
邮箱 :
Honors and Titles:
教育部课程思政示范课“传感器技术及应用”(排6) 2021
北航教学优秀奖二等奖 2021
北航优秀教学成果奖一等奖(排12) 2021
北航优秀教学成果二等奖(排4)
北航优秀教学成果奖三等奖(排3) 2020
The Last Update Time : ..