扫描手机二维码

欢迎您的访问
您是第 位访客

开通时间:..

最后更新时间:..

  • 屈晓磊

    的个人主页 http://shi.buaa.edu.cn/quxiaolei/zh_CN/index.htm

  •   博士生导师   硕士生导师
  • 主要任职:副教授
论文 当前位置: 中文主页 >> 论文
A VGG attention vision transformer network for benign and malignant classification of breast ultrasound images
点击次数:
影响因子:4.506
DOI码:10.1002/mp.15852
发表刊物:Medical Physics
摘要:Purpose: Breast cancer is the most commonly occurring cancer worldwide. The ultrasound reflectivity imaging technique can be used to obtain breast ultrasound (BUS) images, which can be used to classify benign and malignant tumors. However, the classification is subjective and dependent on the experience and skill of operators and doctors. The automatic classification method can assist doctors and improve the objectivity, but current convolution neural network (CNN) is not good at learning global features and vision transform (ViT) is not good at extraction local features. In this study, we proposed an VGG attention vision transformer (VGGA-ViT) network to overcome their disadvantages. Methods: In the proposed method, we used a convolutional neural network (CNN) module to extract the local features and employed a vision transformer (ViT) module to learn the global relationship between different regions and enhance the relevant local features. The CNN module was named the VGG attention (VGGA) module. It was composed of a visual geometry group (VGG) backbone, a feature extraction fully connected layer, and a squeeze-and-excitation (SE) block. Both the VGG backbone and the ViT module were pre-trained on the ImageNet dataset and re-trained using BUS samples in this study. Two BUS datasets were employed for validation. Results: Cross-validation was conducted on two BUS datasets. For the Dataset A, the proposed VGGA-ViT network achieved high accuracy (88.71±1.55%), recall (90.73±1.57%), specificity (85.58±3.35%), precision (90.77±1.98%), F1 score (90.73±1.24%), and Matthews correlation coefficient (MCC) (76.34±3.29%), which were better than those of all compared previous networks in this study. The Dataset B was used as a separate test set, the test results showed that the VGGA-ViT had highest accuracy (81.72±2.99%), recall (64.45±2.96%), specificity (90.28± 3.51%), precision (77.08±7.21%), F1 score (70.11±4.25%), and MCC (57.64±6.88%). Conclusions: In this study, we proposed the VGGA-ViT for the BUS classification, which was good at learning both local and global features. The proposed network achieved higher accuracy than the compared previous methods.
是否译文:否
发表时间:2022-06-22
收录刊物:SCI
版权所有 2014-2022 北京航空航天大学  京ICP备05004617-3  文保网安备案号1101080018
地址:北京市海淀区学院路37号  邮编:100191  电话:82317114