Qu Xiaolei
Personal Homepage
Paper
Current location: Home >> Paper
A VGG attention vision transformer network for benign and malignant classification of breast ultrasound images
Hits:

Impact Factor:4.506

DOI number:10.1002/mp.15852

Journal:Medical Physics

Abstract:Purpose: Breast cancer is the most commonly occurring cancer worldwide. The ultrasound reflectivity imaging technique can be used to obtain breast ultrasound (BUS) images, which can be used to classify benign and malignant tumors. However, the classification is subjective and dependent on the experience and skill of operators and doctors. The automatic classification method can assist doctors and improve the objectivity, but current convolution neural network (CNN) is not good at learning global features and vision transform (ViT) is not good at extraction local features. In this study, we proposed an VGG attention vision transformer (VGGA-ViT) network to overcome their disadvantages. Methods: In the proposed method, we used a convolutional neural network (CNN) module to extract the local features and employed a vision transformer (ViT) module to learn the global relationship between different regions and enhance the relevant local features. The CNN module was named the VGG attention (VGGA) module. It was composed of a visual geometry group (VGG) backbone, a feature extraction fully connected layer, and a squeeze-and-excitation (SE) block. Both the VGG backbone and the ViT module were pre-trained on the ImageNet dataset and re-trained using BUS samples in this study. Two BUS datasets were employed for validation. Results: Cross-validation was conducted on two BUS datasets. For the Dataset A, the proposed VGGA-ViT network achieved high accuracy (88.71±1.55%), recall (90.73±1.57%), specificity (85.58±3.35%), precision (90.77±1.98%), F1 score (90.73±1.24%), and Matthews correlation coefficient (MCC) (76.34±3.29%), which were better than those of all compared previous networks in this study. The Dataset B was used as a separate test set, the test results showed that the VGGA-ViT had highest accuracy (81.72±2.99%), recall (64.45±2.96%), specificity (90.28± 3.51%), precision (77.08±7.21%), F1 score (70.11±4.25%), and MCC (57.64±6.88%). Conclusions: In this study, we proposed the VGGA-ViT for the BUS classification, which was good at learning both local and global features. The proposed network achieved higher accuracy than the compared previous methods.

Translation or Not:no

Date of Publication:2022-06-22

Included Journals:SCI

Attachments:

Personal information

Supervisor of Doctorate Candidates
Supervisor of Master's Candidates

E-Mail:

Date of Employment:2017-05-01

School/Department:School of Instrumentation and Optoelectronic Engineering

Administrative Position:Vice Dean of Department

Business Address:New building B504, School of Instrumentation and Optoelectronic Engineering, Beihang University

Gender:Male

Contact Information:quxiaolei@gmail.com

Status:Employed

Academic Titles:Associate professor

Alma Mater:the University of Tokyo

Discipline:Instrumentation Science and Technology

邮箱 :

Honors and Titles:

教育部课程思政示范课“传感器技术及应用”(排6)  2021

北航教学优秀奖二等奖  2021

北航优秀教学成果奖一等奖(排12)  2021

北航优秀教学成果二等奖(排4)  

北航优秀教学成果奖三等奖(排3)  2020

You are visitors

The Last Update Time : ..


Copyright © 2022 Beihang University. All rights reserved.
Address: 37 Xueyuan Road, Haidian District, Beijing, P.R. China, 100191.
Tel: +86-10-82317114

MOBILE Version