Evaluasi Strategi Fine-Tuning pada ConvNeXt dan Swin Transformer untuk Klasifikasi Kanker Kulit
Abstract
Skin cancer is one of the diseases whose prevalence continues to increase every year, especially in areas with high exposure to ultraviolet (UV) rays. The main challenge in diagnosing skin cancer lies in the visual similarity between benign and malignant lesions, which often leads to misdiagnosis even by experienced medical personnel. The development of deep learning technology has made significant progress in medical image classification through a transfer learning approach. This study aims to compare the performance of two architectures from Transformer and CNN, namely Swin Transformer and ConvNeXt, in the task of classifying two class benign and malignant skin cancer images. Both models use pretrained from ImageNet and are applied with three different fine-tuning strategies, namely Linear Probe (LP), Full Fine-Tuning (FT), and a combination of the two previous strategies (LP-FT). The dataset used is the ISIC Archive Dataset with an 80:20 data split for training and validation, consisting of 3.297 images divided into two classes, with 1800 benign images and 1.497 malignant images. The evaluation was performed using the accuracy, precision, recall, and F1-score metrics. Swin Transformer with the LP-FT strategy achieved the best performance, with an accuracy of 92,27%, precision of 92,24%, recall of 92,17%, and an F1-score of 92,20%. These findings indicate that the two-stage fine-tuning approach can improve model stability and generalization, as well as contribute to the development of a more accurate artificial intelligence based skin cancer diagnosis system.
Downloads
References
S. Gondhowiardjo et al., “Five-Year Cancer Epidemiology at the National Referral Hospital: Hospital-Based Cancer Registry Data in Indonesia,” JCO Glob. Oncol., no. 7, pp. 190–203, 2021, doi: 10.1200/go.20.00155.
Y. L. Vechtomova, T. A. Telegina, A. A. Buglak, and M. S. Kritsky, “Uv Radiation in Dna Damage and Repair Involving Dna-photolyases and Cryptochromes,” Biomedicines, vol. 9, no. 11, pp. 1–13, 2021, doi: 10.3390/biomedicines9111564.
L. Zhou, Y. Zhong, L. Han, Y. Xie, and M. Wan, “Global, Regional, and National Trends in the Burden of Melanoma and Non-melanoma Skin Cancer: Insights From the Global Burden of Disease Study 1990–2021,” Sci. Rep., vol. 15, no. 1, p. 5996, Feb. 2025, doi: 10.1038/s41598-025-90485-3.
L. T. Simanjuntak, “Genetic Predisposition to Malignant Melanoma in the Population of Batam, Indonesia: A Case-Control Study,” Sci. J. Dermatology Venereol., vol. 1, no. 2, pp. 87–99, Sep. 2023, doi: 10.59345/sjdv.v1i2.57.
A. S. Al-Waisy, S. Al-Fahdawi, M. I. Khalaf, M. A. Mohammed, B. Al-Attar, and M. N. Al-Andoli, “A Deep Learning Framework for Automated Early Diagnosis and Classification of Skin Cancer Lesions in Dermoscopy Images,” Sci. Rep., vol. 15, no. 1, pp. 1–20, 2025, doi: 10.1038/s41598-025-15655-9.
A. Bello, S. C. Ng, and M. F. Leung, “Skin Cancer Classification Using Fine-Tuned Transfer Learning of DENSENET-121,” Appl. Sci., vol. 14, no. 17, 2024, doi: 10.3390/app14177707.
G. Papanastasiou, N. Dikaios, J. Huang, C. Wang, and G. Yang, “Is Attention all You Need in Medical Image Analysis? A Review,” IEEE J. Biomed. Heal. Informatics, vol. 28, no. 3, pp. 1398–1411, 2024, doi: 10.1109/JBHI.2023.3348436.
A. Dosovitskiy et al., “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale,” 2021, arXiv. doi: 10.48550/arXiv.2010.11929.
Z. Liu et al., “Swin Transformer: Hierarchical Vision Transformer using Shifted Windows,” Proc. IEEE Int. Conf. Comput. Vis., pp. 9992–10002, 2021, doi: 10.1109/ICCV48922.2021.00986.
Z. Liu, H. Mao, C. Y. Wu, C. Feichtenhofer, T. Darrell, and S. Xie, “A ConvNet for the 2020s,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., vol. 2022-June, pp. 11966–11976, 2022, doi: 10.1109/CVPR52688.2022.01167.
M. Ramadhan and J. Zeniarja, “Implementation of Deep Transfer Learning and Explainable AI in Skin Cancer Classification,” SISTEMASI, vol. 14, p. 2266, Sep. 2025, doi: 10.32520/stmsi.v14i5.5425.
A. Newaz, A. U. R. Adib, R. Sahil, and M. Mehzad, “An End-to-End Deep Learning Framework for Arsenicosis Diagnosis Using Mobile-Captured Skin Images,” 2025, arXiv. doi: 10.48550/arXiv.2509.08780.
G. H. Dagnaw, M. El Mouhtadi, and M. Mustapha, “Skin Cancer Classification Using Vision Transformers and Explainable Artificial Intelligence,” J. Med. Artif. Intell., vol. 7, no. January, pp. 14–14, Jun. 2024, doi: 10.21037/jmai-24-6.
E. A. Taufik, A. Khondoker, A. F. Parsa, and S. A. M. Mostafa, “Visual Bias and Interpretability in Deep Learning for Dermatological Image Analysis,” in 2025 4th International Conference on Image Processing and Media Computing (ICIPMC), IEEE, Jun. 2025, pp. 45–49. doi: 10.1109/ICIPMC66319.2025.11170678.
J. Vieira, F. Mendonça, and F. Morgado-Dias, “Deep Learning Approaches for Skin Lesion Detection,” Electronics, vol. 14, no. 14, p. 2785, 2025, doi: 10.3390/electronics14142785.
A. Davila, J. Colan, and Y. Hasegawa, “Comparison of Fine-tuning Strategies for Transfer Learning in Medical Image Classification,” Image Vis. Comput., vol. 146, p. 105012, Jun. 2024, doi: 10.1016/j.imavis.2024.105012.
A. Kumar, A. Raghunathan, R. Jones, T. Ma, and P. Liang, “Fine-Tuning Can Distort Pretrained Features and Underperform Out-of-Distribution,” ICLR 2022 - 10th Int. Conf. Learn. Represent., vol. 10, no. Id, pp. 1–54, 2022, doi: 10.48550/arXiv.2202.10054.
H. E. Kim, A. Cosa-Linan, N. Santhanam, M. Jannesari, M. E. Maros, and T. Ganslandt, “Transfer learning for medical image classification: a literature review,” BMC Med. Imaging, vol. 22, no. 1, p. 69, Dec. 2022, doi: 10.1186/s12880-022-00793-7.
L. Alzubaidi et al., “Novel Transfer Learning Approach for Medical Imaging with Limited Labeled Data,” Cancers (Basel)., vol. 13, no. 7, p. 1590, Mar. 2021, doi: 10.3390/cancers13071590.
S. Sathyanarayanan, “Confusion Matrix-Based Performance Evaluation Metrics,” African J. Biomed. Res., pp. 4023–4031, 2024, doi: 10.53555/AJBR.v27i4S.4345.
Ž. Ð. Vujovic, “Classification Model Evaluation Metrics,” Int. J. Adv. Comput. Sci. Appl., vol. 12, no. 6, 2021, doi: 10.14569/IJACSA.2021.0120670.
D. R. I. M. Setiadi, K. Nugroho, A. R. Muslikh, S. W. Iriananda, and A. A. Ojugo, “Integrating SMOTE-Tomek and Fusion Learning with XGBoost Meta-Learner for Robust Diabetes Recognition,” J. Futur. Artif. Intell. Technol., vol. 1, no. 1, pp. 23–38, 2024, doi: 10.62411/faith.2024-11.
R. Barinov, V. Gai, G. Kuznetsov, and V. Golubenko, “Automatic Evaluation of Neural Network Training Results,” Computers, vol. 12, no. 2, p. 26, 2023, doi: 10.3390/computers12020026.
Bila bermanfaat silahkan share artikel ini
Berikan Komentar Anda terhadap artikel Evaluasi Strategi Fine-Tuning pada ConvNeXt dan Swin Transformer untuk Klasifikasi Kanker Kulit
Pages: 2240-2249
Copyright (c) 2026 Ahmad Bintang Saputra, Sindhu Rakasiwi

This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under Creative Commons Attribution 4.0 International License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (Refer to The Effect of Open Access).





















