Please use this identifier to cite or link to this item: http://localhost:8080/xmlui/handle/123456789/3380
Title: BOOTSTRAPPING OF FINE-TUNED SEGMENTATION AND CLASSIFICATION NETWORK FOR EPIDERMIS DISORDER CATEGORIZATION
Authors: Kalaivani A
Karpagavalli S
Issue Date: 25-Jul-2023
Publisher: Springer Link
Abstract: As all people have been affected by various skin related illnesses, categorization of skin disorders has become prominent in recent healthcare system. To identify and categorize skin related syndromes, many transfer learning frameworks were used. Amongst, a Fine-tuned Segmentation and Classification Network (F-SegClassNet) achieved better efficacy by using the novel unified loss function. Nonetheless, it was not apt for the datasets that lack in training images. Hence in this article, Bootstrapping of F-SegClassNet (BF-SegClassNet) model is proposed which solves the imbalanced images in the training set via generating the group of pseudo balanced training batches relying on the properties of the considered skin image dataset. This model fits the distinct abilities of Deep Convolutional Neural Network (DCNN) classifier so that it is highly useful for classifying the skin disorder image dataset with a highly imbalanced image data distribution. According to the Bootstrpping, better tradeoff between simple and complex image samples is realized to make a network model that is suitable for automatic skin disorders classification. In this model, statistics across the complete training set is calculated and a new subset is produced that retains the most essential image samples. So, the skin images are segmented and categorized by this new model to identify the varieties of epidermis infections. At last, the testing outcomes exhibits BF-SegClassNet-model accomplishes the mean accuracy with 96.14% for HAM dataset which is compared to state-of-the-art models.
URI: https://link.springer.com/article/10.1007/s11042-023-16255-3
Appears in Collections:National Journals



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.