Keywords

1 Introduction

Burns are amongst the severe global health issues, and as per WHO Burns Reports 2018, an estimated 180,000 persons lose their life because of burns each year [1]. And many more become crippled forever. Just in India alone, more than 1 million get moderately or severely burnt every year [1, 2]. Around 70% burn injuries occur in the most productive age group (25 ± 10 years), and most patients belong to poor socioeconomic strata. Also, the post burn life of burns survivor is never the same - socially, economically, mentally and physically. Timely done burns diagnosis and adequate first aid treatment can check the burns severities and significantly reduce the number of burn death cases. However, there are several challenges such as limited number of expert dermatologists, limited number of burn centres, lack of awareness about the first aid and prevention strategies, use of manual methods in diagnosis, and heavy treatment costs. Despite advancements in burn care, proper burns diagnosis and treatment of burn patients remains a major challenge. Automated methods to give an early assessment of the total body surface area (TBSA) burnt and/or the burns depth can be extremely helpful for better burns diagnosis [2,3,4,5]. Laser Doppler imaging (LDI) has been found as most reliable technique for burn depth assessment but high costs, delays and limited portability limits its real world usage [4, 6, 7]. Considering high penetrations of Smartphones having good resolution cameras and advances in data sciences, visual images based automated burns diagnosis methods can be very useful.

A group of researchers from University of Seville, Spain had been actively contributing in this domain of burn depth assessment methods using color images from around last two decades [3, 8,9,10]. Lately, they presented a psychophysical experiment and multidimensional scaling analysis to determine the physical characteristics that physicians employ to diagnose a burn depth [3]. In this work, they used a k-nearest neighbor classifier on a dataset of 74 images, and the accuracy of 66.2% was achieved in classifying the burn images considered into three burn depths with their approach. Earlier, Wantanajittikul et al. described a support vector machine (SVM) based approach for burn degree assessment [11] but they also considered very less images. Recently, Badea et al. [5] proposed an ensemble method build upon fusing decision from standard classifiers (SVM and Random Forest) and the CNN architecture ResNet [12] for burn severity assessment. For feature extraction, they used primarily the Histogram of Topographical features (HoT). Their proposed system was able to identify light/serious burns with an average precision of 65%.

As the skin architecture varies across different parts of the body [13, 14], and so the burns impact on different body parts skin would be different. Therefore, it is likely that the body part specific visual images based automatic burns diagnosis assessment methods would be more effective than generic visual images based methods. Considering this, we explore this problem of classifying the body part of burn images. To the best of our knowledge, ours is the first attempt to classify burnt body part images. In this work, we consider four different burnt body parts: face, hand, back, and inner arm, and we aim to identify the body-part of the given input burnt image. We then discuss the performance of ResNet50 deep learning architecture based independent and dependent deep learning models in classifying the different burnt body parts images.

2 Data Collection

Using Google Search engine, we gathered burned images of different body parts [15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39]. Out of these total 109 burnt images that we collected, there are 30 face images, 35 hand images, 23 back images and rest are inner forearm images. The average resolution of these images is in the range of (350 − 450) × (300 − 400) pixels. Figures 1, 2 and 3 show some sample burnt images of back, hand and inner forearm.

Fig. 1.
figure 1

Sample burnt images of human back body parts

Fig. 2.
figure 2

Sample burnt images of human hand [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39]

Fig. 3.
figure 3

Sample burnt images of inner forearm [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39]

We also used in one of our models a dataset of 4981 non-burnt skin images of these four different parts: back, face, hand and inner forearm. We used some available datasets for these images [40,41,42,43].

3 Methodology

We use the ResNet50 architecture deep learning models (as shown in Fig. 4) for feature extraction [12]. We freeze all the layers of the ResNet-50 model and replace its output layer by a dense layer with 4 nodes, representing 4 classes. The average pooling method is used for feature extraction. We use Softmax activation function, ADAM optimizer and categorical cross-entropy loss function in our model. We used two variants of this model: M1 - dependent and M2 - independent (considering limited number of burn images), for the classification of the body-part of the input burnt image.

Fig. 4.
figure 4

Block diagram of the ResNet50 architecture used [44]

For dependent model M1, we used leave-one-out cross validation (LOOCV) for classification. Noting the challenge of limited number of burnt images availability, we explored the independent deep learning model and its effectiveness for this classification problem. In the independent model M2, we used a total of 4981 non-burnt skin images of four different parts (face images: 2000 [40], hand images: 2000 [41], back: 81 [42], and inner forearm: 800 images [43]) for the training, a total of 1798 non-burnt skin images for the validation [40,41,42,43], and the same 109 burnt body part images for the testing. To account for class imbalance in training, we normalized the class weights to be inversely proportional to number of images in that class.

4 Experimental Results

For the 109 burnt body parts images, we obtained overall classification accuracy of 90.83% and 93.58% using dependent (M1) and independent (M2) deep learning ResNet-50 model, respectively. The M1 used LOOCV and thus training was performed 109 times. It accurately classified 22 back body part burn images, 27 face images, 31 hand images and 19 inner forearm burnt images.

The model M2 that was trained using no burn images classified in comparison 2 back and 3 forearm burnt image inaccurately, but M2 performed much better in classifying other body-parts burnt images as 29 face and 34 hand burnt images were classified accurately (Tables 2 and 3). Table 1 presents the classification accuracy results and effectiveness of these two models. When we consider the top 2 estimated classes for each of the test image, both the models were found to be more than 99% accurate.

Table 1. Classification accuracy results.
Table 2. Confusion matrix for model M1
Table 3. Confusion matrix for model M2

5 Conclusions and Future Work

We discussed a new problem of classifying burnt body part images that can help in the development of better burns diagnosis methods and we present the effectiveness of some deep learning models in addressing this problem, considering the constraint of limited number of burnt images availability. It is encouraging and interesting to note that a ResNet50 architecture based model, trained using non burnt images, was able to classify with more than 93% accuracy the burnt images of 4 different body parts: face, back, hand, and inner forearm. Further, there is a scope of exploring the robustness and effectiveness of the proposed method using much larger number of burnt images and considering more classes by including other body parts e.g. feet, stomach or chest.