A Self-Supervised Vision Transformer Approach for Dermatological Image Analysis
Main Article Content
Abstract
This paper proposes a novel skin disease classification method, which combines visual transformers (ViT) with self-supervised learning and is validated on the ISIC 2018 dataset. Experimental results show that the method can effectively extract feature representations through feature pre-training through self-supervised contrastive learning, and shows significant advantages over traditional convolutional neural network (CNN) models in multiple indicators such as AUC, precision, recall and F1 score. T-SNE visualization further reveals the obvious clustering characteristics in the feature space, confirming the superiority of the method. Compared with existing mainstream technologies, this method shows higher discrimination ability in handling complex skin lesion classification tasks, while reducing the dependence on large-scale labeled data and enhancing the practicality of the model. Analysis of the loss function curve shows that the method achieves fast and stable convergence during training, highlighting its efficiency and stability. This study verifies the potential of ViT in medical image analysis and provides an efficient solution for the automatic classification of skin diseases.
Article Details

This work is licensed under a Creative Commons Attribution 4.0 International License.
Mind forge Academia also operates under the Creative Commons Licence CC-BY 4.0. This allows for copy and redistribute the material in any medium or format for any purpose, even commercially. The premise is that you must provide appropriate citation information.