Journal of Jilin University(Engineering and Technology Edition) ›› 2025, Vol. 55 ›› Issue (11): 3686-3696.doi: 10.13229/j.cnki.jdxbgxb.20240222

Previous Articles    

Chest X-ray images classification based on multi-scale attention information multiplexing network

Rui-feng ZHANG(),Fang-zhao GUO,Qiang LI()   

  1. School of Microelectronics,Tianjin University,Tianjin 300072,China
  • Received:2024-03-05 Online:2025-11-01 Published:2026-02-03
  • Contact: Qiang LI E-mail:zhangruifeng@tju.edu.cn;liqiang@tju.edu.cn

Abstract:

To address issues such as low recognition of lesion areas in chest X-ray images and the difficulty in accurately capturing the spatial positions of lesions, a multi-scale attention information multiplexing network that helps improve the dassification accuracy of chest X-ray images was proposed in this paper. Firstly, by introducing multiple spatial information multiplexing blocks, the network enhances the positional connections between disease regions on feature maps and across channels; Secondly, through a multi-scale integration attention blocks, the network integrates multi-scale image feature information to automatically capture disease location variations and flexibly focus on key pathological information; Finally, the problem of imbalanced distribution of chest disease samples was alleviated by using an asymmetric shift focus loss function. Multiple experiments on the publicly available datasets ChestX-ray14 and CheXpert have shown that the average area under curve (AUC) value of the proposed network on two datasets reached 0.847 and 0.901 respectively, which is superior the more advanced network models in recent years. This indicates that the network can effectively improve the classification accuracy of chest diseases.

Key words: computer application technology, chest X-ray image classification, spatial information multiplexing, multi-scale attention, asymmetric shift focus loss function

CLC Number: 

  • TP391

Fig.1

MIM-Net model architecture"

Fig.2

Multiple spatial information multiplexing block"

Fig.3

Multi-scale integration attention block"

Fig.4

Number of images of each type in ChestX-ray14 dataset"

Table 1

Number of images of each type in CheXpert dataset"

疾病种类阳性不确定阴性
肺不张29 33329 377165 606
心脏肿大23 0026 597194 717
肺实变12 73023 976187 610
水肿48 90511 571163 840
胸腔积液75 6969 419139 201

Fig.5

AUC scatter plots of different λ combinations"

Fig.6

ROC curves of 14 diseases on the ChestX-ray14 datast"

Table 2

Comparison of the classification accuracy of each network using the ChestX-ray14 dataset"

疾病种类DCNNConsultNetDeformab-CDAM-DA3NetPCSANetCheXGATPCANSSGEMIM-Net
平均AUC0.7450.8220.8400.8260.8250.8270.8240.8300.847
肺不张0.7000.7850.8200.7790.8070.7870.7850.7920.826
心脏肿大0.8100.8990.9120.8950.9100.8790.8970.8920.919
积液0.7590.8350.8900.8360.8790.8370.8370.8400.887
渗透0.6610.6990.7140.7100.6980.6990.7060.7140.715
肿块0.6930.8380.8650.8340.8240.8390.8340.8480.869
肺结节0.6690.7750.7720.7770.7500.7930.7860.8120.784
肺炎0.6580.7380.7620.7370.7500.7410.7300.7330.801
气胸0.7990.8710.9030.8780.8500.8790.8710.8850.892
肺实变0.7030.7630.8100.7590.8020.7550.7630.7530.809
水肿0.8050.8500.8960.8550.8880.8510.8540.8480.900
肺气肿0.8330.9240.9140.9330.8900.9450.9210.9480.924
纤维化0.7860.8310.8080.8380.8120.8420.8170.8270.824
胸膜增厚0.6840.7760.8150.7910.7680.7940.7910.7950.822
疝气0.8720.9220.8760.9380.9150.9310.9430.9320.889

Table 3

Comparison of the classification accuracy of each network on the CheXpert dataset"

疾病类型

Ensemble

(U-Ones)

ConsultNet

(U-Ones)

PCAN

(U-Ones)

MIM-Net

(U-Ones)

Ensemble

(U-Zeros)

ConsultNet

(U-Zeros)

DCNN

(U-Zeros)

MIM-Net

(U-Zeros)

肺不张0.8580.8470.8480.8590.8110.8040.7450.842
心脏肿大0.8320.8680.8650.8850.8400.8740.8130.873
肺实变0.8990.9230.9080.9040.9320.9400.8820.901
水肿0.9410.9240.9120.9350.9290.8940.9210.928
积液0.9340.9260.9400.9240.9310.9230.9300.923
平均值0.8930.8980.8950.9010.8890.8890.8580.893

Table 4

Comparison results of the ablation experiment on MSIM and MIA module"

MSIM××
MIA××
肺不张0.8220.8110.8240.826
心脏肿大0.9110.9070.9180.919
积液0.8770.8920.8850.887
渗透0.7140.7180.7150.714
肿块0.8650.8580.8660.869
肺结节0.7790.7910.7860.784
肺炎0.7650.7710.7990.801
气胸0.8520.8790.8760.892
肺实变0.8170.8010.8050.809
水肿0.8950.8910.8970.900
肺气肿0.9170.9150.9260.924
纤维化0.8030.8170.8140.824
胸膜增厚0.8190.8080.8230.822
疝气0.8920.8850.8810.889
平均AUC0.8380.8390.8440.847

Table 5

Influence of loss function on the results"

损失函数平均AUC
ChestX-ray14CheXpert
交叉熵损失函数0.8380.889
焦点损失函数0.8430.893
非对称移位焦点损失函数0.8470.901

Fig.7

Lesion area annotation map and visual heat map"

[1] Wei X L, Li W, Zhang M M, et al. Medical hyperspectral image classification based on end-to-end fusion deep neural network[J]. IEEE Transactions on Instrumentation and Measurement, 2019, 68(11): 4481-4492.
[2] 刘桂霞, 田郁欣, 王涛, 等. 基于双输入3D卷积神经网络的胰腺分割算法[J]. 吉林大学学报: 工学版, 2023, 53(12): 3565-3572.
Liu Gui-xia, Tian Yu-xin, Wang Tao, et al. Pancreas segmentation algorithm based on dual input 3D convolutional neural network[J]. Journal of Jilin University (Engineering and Technology Edition), 2023, 53(12): 3565-3572.
[3] 王雪, 李占山, 吕颖达. 基于多尺度感知和语义适配的医学图像分割算法[J]. 吉林大学学报: 工学版, 2022, 52(3): 640-647.
Wang Xue, Li Zhan-shan, Ying-da Lyu. Medical image segmentation based on multi-scale context-aware and semantic adaptor[J]. Journal of Jilin University (Engineering and Technology Edition), 2022, 52(3): 640-647.
[4] Wang X S, Peng Y F, Lu L, et al. ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases[C]∥IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE, USA, 2017: 3462-3471.
[5] Krizhevsky A, Sutskever I, Hinton E G. ImageNet classification with deep convolutional neural networks[J]. Communications of the ACM, 2017, 60(6): 84-90.
[6] He K M, Zhang X Y, Ren S Q, et al. Deep residual learning for image recognition[C]∥IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Piscataway, NJ: IEEE, USA, 2016: 770-778.
[7] Irvin J, Rajpurkar P, Ko M, et al. CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison[C]∥Proceedings of the AAAI Conference on Artificial Intelligence Washington, DC: AAAI Press, 2019: 590-597.
[8] Jiang X B, Zhu Y, Cai G, et al. MXT: A new variant of pyramid vision transformer for multi-label chest x-ray image classification[J]. Cognitive Computation, 2022, 14(4): 1362-1377.
[9] Wang W H, Xie E Z, Li X, et al. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions[C]∥2021 IEEE/CVF International Conference on Computer Vision. Piscataway, NJ: IEEE, 2021: 548-558.
[10] 胡锦波, 聂为之, 宋丹, 等. 可形变Transformer辅助的胸部X光影像疾病诊断模型[J]. 浙江大学学报:工学版, 2023, 57(10): 1923-1932.
Hu Jin-bo, Nie Wei-zhi, Song Dan, et al. Chest X-ray imaging disease diagnosis model assisted by deformable transformer[J]. Journal of Zhejiang University (Engineering Science), 2023, 57(10): 1923-1932.
[11] Wang H Y, Wang S S, Qin Z B, et al. Triple attention learning for classification of 14 thoracic diseases using chest radiography[J]. Medical Image Analysis, 2021, 67: 101846.
[12] Zhu X F, Pang S M, Zhang X X, et al. PCAN: Pixel-wise classification and attention network for thoracic disease classification and weakly supervised localization[J]. Computerized Medical Imaging and Graphics, 2022, 102: 102137.
[13] Chen K, Wang X Q, Zhang S W. Thorax disease classification based on pyramidal convolution shuffle attention neural network[J]. IEEE Access, 2022, 10: 85571-85581.
[14] Lu Z C, Deb K, Boddeti V N. MUXConv: Information multiplexing in convolutional neural networks[J]. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2020, 6: 12041-12050.
[15] Woo S H, Park J, Lee J Y, et al. CBAM: Convolutional block attention module[C]∥Computer Vision-ECCV 2018. Cham, Switzerland: Springer Mature Switzerland AG, 2018: 3-19.
[16] Ridnik T, Ben B E, Zamir N, et al. Asymmetric loss for multi-label classification[C]∥IEEE/CVF International Conference on Computer Vision. Piscataway, NJ: IEEE, 2021: 82-91.
[17] Guan Q J, Huang Y P, Luo Y W, et al. Discriminative feature learning for thorax disease classification in chest X-ray images[J]. IEEE Transactions on Image Processing, 2021, 30: 2476-2487.
[18] Lee Y W, Huang S K, Chang R F. CheXGAT: A disease correlation-aware network for thorax disease diagnosis from chest X-ray images[J]. Artificial Intelligence in Medicine, 2022, 132: 102382.
[19] Chen B Z, Zhang Z, Li Y J, et al. Multi-label chest X-ray image classification via semantic similarity graph embedding[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(4): 2455-2468.
[20] Pham H, Le T, Tran D, et al. Interpreting chest X-rays via CNNs that exploit hierarchical disease dependencies and uncertainty labels[J]. Neurocomputing, 2021, 437: 186-194.
[21] Selvaraju R, Cogswell M, Das A, et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization[C]∥IEEE International Conference on Computer Vision. Piscataway, NJ: IEEE, 2017: 618-626.
[1] Yu-fei ZHANG,Li-min WANG,Jian-ping ZHAO,Zhi-yao JIA,Ming-yang LI. Robot inverse kinematics solution based on center selection battle royale optimization algorithm [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(8): 2703-2710.
[2] Wen-hui LI,Chen YANG. Few-shot remote sensing image classification based on contrastive learning text perception [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(7): 2393-2401.
[3] Xiang-jiu CHE,Liang LI. Graph similarity measurement algorithm combining global and local fine-grained features [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(7): 2365-2371.
[4] Jian WANG,Chen-wei JIA. Trajectory prediction model for intelligent connected vehicle [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(6): 1963-1972.
[5] Feng-feng ZHOU,Zhe GUO,Yu-si FAN. Feature representation algorithm for imbalanced classification of multi⁃omics cancer data [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(6): 2089-2096.
[6] Xiang-jiu CHE,Yu-peng SUN. Graph node classification algorithm based on similarity random walk aggregation [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(6): 2069-2075.
[7] Xiang-jiu CHE,Yu-ning WU,Quan-le LIU. A weighted isomorphic graph classification algorithm based on causal feature learning [J]. Journal of Jilin University(Engineering and Technology Edition), 2025, 55(2): 681-686.
[8] Li-ming LIANG,Long-song ZHOU,Jiang YIN,Xiao-qi SHENG. Fusion multi-scale Transformer skin lesion segmentation algorithm [J]. Journal of Jilin University(Engineering and Technology Edition), 2024, 54(4): 1086-1098.
[9] Dondrub LHAKPA,Duoji ZHAXI,Jie ZHU. Tibetan text normalization method [J]. Journal of Jilin University(Engineering and Technology Edition), 2024, 54(12): 3577-3588.
[10] Yu-xin YE,Luo-jia XIA,Ming-hui SUN. Gesture input method based on transparent keyboard in augmented reality environment [J]. Journal of Jilin University(Engineering and Technology Edition), 2024, 54(11): 3274-3282.
[11] Na CHE,Yi-ming ZHU,Jian ZHAO,Lei SUN,Li-juan SHI,Xian-wei ZENG. Connectionism based audio-visual speech recognition method [J]. Journal of Jilin University(Engineering and Technology Edition), 2024, 54(10): 2984-2993.
[12] Ya-hui ZHAO,Fei-yu LI,Rong-yi CUI,Guo-zhe JIN,Zhen-guo ZHANG,De LI,Xiao-feng JIN. Korean⁃Chinese translation quality estimation based on cross⁃lingual pretraining model [J]. Journal of Jilin University(Engineering and Technology Edition), 2023, 53(8): 2371-2379.
[13] Shan XUE,Ya-liang ZHANG,Qiong-ying LYU,Guo-hua CAO. Anti⁃unmanned aerial vehicle system object detection algorithm under complex background [J]. Journal of Jilin University(Engineering and Technology Edition), 2023, 53(3): 891-901.
[14] Jin-Zhen Liu,Guo-Hui Gao,Hui Xiong. Multi⁃scale attention network for brain tissue segmentation [J]. Journal of Jilin University(Engineering and Technology Edition), 2023, 53(2): 576-583.
[15] Zhen WANG,Xiao-han YANG,Nan-nan WU,Guo-kun LI,Chuang FENG. Ordinal cross entropy Hashing based on generative adversarial network [J]. Journal of Jilin University(Engineering and Technology Edition), 2023, 53(12): 3536-3546.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!