Integrating Self-Supervised Learning with Nonlinear Classifiers in Lightweight Swin Transformer for X-Ray Image Classification

  • Tri-Thuc Vo College of Information Technology, Can Tho University, 92000-Cantho, Vietnam
  • Thanh-Nghi Do UMI UMMISCO 209 (IRD/UPMC), Sorbonne University, Pierre and Marie Curie University- Paris 6, France
Keywords: Self-supervised learning, X-ray image, swin transformer, multi-class classification

Abstract

 In this paper, we present a new approach about the integration Self-Supervised Learning with nonlinear
Classifiers in Lightweight Swin Transformer (SSLnC-LSwinT) for improving performance of X-ray image classification. Our approach leverages unlabeled data to address the issue of labeled data scarcity in the medical field by using self supervised learning (SSL) to extract features. One of our key contributions is the introduction of the Lightweight SwinT architecture, a more lightweight variant of SwinT, designed to enhance computational efficiency, reduce model complexity, and shorten training time. To further improve  classification efficiency, we propose the integration of a nonlinear classifier instead of a linear classifier in Lightweight SwinT. The experimental results underscore our contributions, demonstrating significant reductions in model training time and notable improvements in classification performance. Our proposed method, which integrates SSL based on LSwinT with a nonlinear LightGBM classifier, achieves an accuracy of up to 87%, improving by 1.8% over the non-LightGBM SwinT version and reducing training time significantly (3:23:00 vs. 7:37:29) compared to the original SwinT architecture.

Author Biographies

Tri-Thuc Vo, College of Information Technology, Can Tho University, 92000-Cantho, Vietnam

Tri-Thuc Vo received the B.Eng. degree in Software Engineering from the Cantho University, Vietnam, in 2011. He received his MSc. degree in informatics from the University of Brest, France, in 2018. He is currently a lecturer at the College of Information Technology, Cantho University, Vietnam. His research interests include
medical data analysis and machine learning.

Thanh-Nghi Do, UMI UMMISCO 209 (IRD/UPMC), Sorbonne University, Pierre and Marie Curie University- Paris 6, France

Thanh-Nghi Do received his PhD. degree in informatics from the University of Nantes, France, in 2004. He is currently an associate professor at the College of Information Technology, Cantho University, Vietnam. He is also an associate researcher at UMIUMMISCO209(IRD/UPMC),Sorbonne University, and the Pierre and Marie Curie University, France. His research interests include data mining with support vector machines, kernel-based methods, decision tree algorithms, ensemble-based learning, and information visualization. He has served on the program committees of international conferences and is a reviewer for journals in his fields of expertise.

References

S. K. Zhou, H. Greenspan, and D. Shen, Deep learning for medical image analysis. Academic Press, 2023.

K. Sailunaz, T. ¨ Ozyer, J. Rokne, and R. Alhajj, “A survey of machine learning-based methods for covid-19 medical image analysis,” Medical & Biological Engineering & Computing, vol. 61, no. 6, pp. 1257–1297, 2023.

A. Dosovitskiy et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” 2021. [Online]. Available: https://arxiv.org/abs/2010.11929

Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” 2021. [Online]. Available: https://arxiv.org/abs/2103.14030

K. He, C. Gan, Z. Li, I. Rekik, Z. Yin, W. Ji, Y. Gao, Q. Wang, J. Zhang, and D. Shen, “Transformers in medical

image analysis,” Intelligent Medicine, vol. 3, no. 1, pp. 59 78, 2023.

K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, “Momentum contrast for unsupervised visual representation learning,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 9729–9738.

X. Chen, H. Fan, R. Girshick, and K. He, “Improved base lines with momentum contrastive learning,” arXiv preprint arXiv:2003.04297, 2020.

X. Chen, S. Xie, and K. He, “An empirical study of training self-supervised vision transformers,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 9640–9649.

Z. Xie, Y. Lin, Z. Yao, Z. Zhang, Q. Dai, Y. Cao, and H. Hu, “Self-supervised learning with swin transformers,”

[Online]. Available: https://arxiv.org/abs/2105.04553

S.-C. Huang, A. Pareek, M. Jensen, M. P. Lungren, S. Yeung, and A. S. Chaudhari, “Self-supervised learning for medical image classification: a systematic review and implementation guidelines,” NPJ Digital Medicine, vol. 6, no. 1, p. 74, 2023.

M. Caron, H. Touvron, I. Misra, H. Jégou, J. Mairal, P. Bojanowski, and A. Joulin, “Emerging properties in

self-supervised vision transformers,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 9650–9660.

H. Sowrirajan, J. Yang, A. Y. Ng, and P. Rajpurkar, “Moco pretraining improves representation and transferability of chest x-ray models,” in Medical Imaging with Deep Learning. PMLR, 2021, pp. 728–744.

K. Cho et al., “Chess: Chest x-ray pre-trained model via self supervised contrastive learning,” Journal of Digital Imaging, vol. 36, no. 3, pp. 902–910, 2023.

P. Rajpurkar and et al., “Deep learning for chest radiograph diagnosis: A retrospective comparison of the chexnext algorithm to practicing radiologists,” PLoS medicine, vol. 15, no. 11, p. e1002686, 2018.

K.-C. Chen and et al., “Diagnosis of common pulmonary diseases in children by x-ray images and deep learning,” Scientific reports, vol. 10, no. 1, p. 17374, 2020.

Y. Jin, H. Lu, W. Zhu, and W. Huo, “Deep learning based classification of multi-label chest x-ray images via dual weighted metric loss,” Computers in Biology and Medicine, vol. 157, p. 106683, 2023.

M. Nahiduzzaman and et al., “Parallel cnn-elm: A multiclass classification of chest x-ray images to identify seventeen lung diseases including covid-19,” Expert Systems with Applications, vol. 229, p. 120528, 2023.

M. Kaya and M. Eris, “D3senet: A hybrid deep feature extraction network for covid-19 classification using chest x-ray images,” Biomedical signal processing and control, vol. 82, p. 104559, 2023.

T.-T. Vo and T.-N. Do, “Improving chest x-ray image classification via integration of self-supervised learning and machine learning algorithms.” Journal of Information & Communication Convergence Engineering, vol. 22, no. 2, 2024.

Z. Ullah, M. Usman, and J. Gwak, “Mtss-aae: Multi-task semi-supervised adversarial autoencoding for covid-19 detection based on chest x-ray images,” Expert Systems with Applications, vol. 216, p. 119475, 2023.

S. U. Amin, S. Taj, A. Hussain, and S. Seo, “An automated chest x-ray analysis for covid-19, tuberculosis, and pneumonia employing ensemble learning approach,” Biomedical Signal Processing and Control, vol. 87, p. 105408, 2024.

S. Taslimi, S. Taslimi, N. Fathi, M. Salehi, and M. H. Rohban, “Swinchex: Multi-label classification on chest x-ray images with transformers,” arXiv preprint arXiv:2206.04246, 2022. [Online]. Available: https://arxiv.org/abs/2206.04246

G. I. Okolo, S. Katsigiannis, and N. Ramzan, “Ievit: An enhanced vision transformer architecture for chest x-ray image classification,” Computer Methods and Programs in Biomedicine, vol. 226, p. 107141, 2022.

J. Ko, S. Park, and H. G. Woo, “Optimization of vision transformer-based detection of lung diseases from chest x-ray images,” BMC Medical Informatics and Decision Making, vol. 24, no. 1, p. 191, 2024.

P. Komorowski, H. Baniecki, and P. Biecek, “Towards evaluating explanations of vision transformers for medical imaging,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 3726 3732.

L. Huang, J. Ma, H. Yang, and Y. Wang, “Research and implementation of multi-disease diagnosis on chest x-ray based on vision transformer,” Quantitative Imaging in Medicine and Surgery, vol. 14, no. 3, p. 2539, 2024.

C. C. Ukwuoma et al., “A hybrid explainable ensemble transformer encoder for pneumonia identification from chest x-ray images,” Journal of Advanced Research, vol. 48, pp. 191–211, 2023.

B. VanBerlo, J. Hoey, and A. Wong, “A survey of the impact of self-supervised pretraining for diagnostic tasks

in medical x-ray, ct, mri, and ultrasound,” BMC Medical Imaging, vol. 24, no. 1, p. 79, 2024.

E. Tiu, E. Talius, P. Patel, C. P. Langlotz, A. Y. Ng, and P. Rajpurkar, “Expert-level detection of pathologies from

unannotated chest x-ray images via self-supervised learning,” Nature Biomedical Engineering, vol. 6, no. 12, pp. 1399 1406, 2022.

M. Shakouri, F. Iranmanesh, and M. Eftekhari, “Dino-cxr: A self supervised method based on vision transformer for chest x-ray classification,” in International Symposium on Visual Computing. Springer, 2023, pp. 320–331.

J. Yao, X. Wang, Y. Song, H. Zhao, J. Ma, Y. Chen, W. Liu, and B. Wang, “Eva-x: A foundation model for

general chest x-ray analysis with self-supervised learning,” 2024. [Online]. Available: https://arxiv.org/abs/2405.05237

G. Li, R. Togo, T. Ogawa, and M. Haseyama, “Covid-19 detection based on self-supervised transfer learning using chest x-ray images,” International Journal of Computer Assisted Radiology and Surgery, vol. 18, no. 4, pp. 715 722, 2023.

J.-B. Grill et al., “Bootstrap your own latent: A new approach to self-supervised learning,” 2020. [Online].

Available: https://arxiv.org/abs/2006.07733

G. Ke, Q. Meng, T. Finley, T. Wang, W. Chen, W. Ma, Q. Ye, and T.-Y. Liu, “Lightgbm: A highly efficient gradient

boosting decision tree,” Advances in neural information processing systems, vol. 30, 2017.

T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting system,” in Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, 2016, pp. 785–794.

L. Prokhorenkova, G. Gusev, A. Vorobev, A. V. Dorogush, and A. Gulin, “Catboost: unbiased boosting with categorical features,” Advances in neural information processing systems, vol. 31, 2018.

F. Pedregosa et al., “Scikit-learn: Machine learning in python,” the Journal of machine Learning research, vol. 12, pp. 2825–2830, 2011.

A. Paszke et al., “Pytorch: An imperative style, high performance deep learning library,” 2019. [Online].

Available: https://arxiv.org/abs/1912.01703

Published
2024-10-27