A CNN BASED MODEL FOR VENOMOUS AND NON-VENOMOUS SNAKE CLASSIFICATION - DIVA PORTAL
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
A CNN Based Model for Venomous and Non-venomous Snake Classification Nagifa Ilma Progga1(B) , Noortaz Rezoana1(B) , Mohammad Shahadat Hossain1 , Raihan Ul Islam2 , and Karl Andersson2 1 Department of Computer Science and Engineering, University of Chittagong, Chittagong, Bangladesh hossain ms@cu.ac.bd 2 Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Skellefteå, Sweden {raihan.ul.islam,karl.andersson}@ltu.se Abstract. Snakes are curved, limbless, warm blooded reptiles of the phylum serpents. Any characteristics, including head form, body shape, physical appearance, texture of skin and eye structure, might be used to individually identify nonvenomous and venomous snakes, that are not usual among non-experts peoples. A standard machine learning method- ology has also been used to create an automated categorization of species of snake dependent upon the photograph, in which the characteristics must be manually adjusted. As a result, a Deep convolutional neural network has been proposed in this paper to classify snakes into two categories: venomous and non-venomous. A set of data of 1766 snake pictures is used to implement seven Neural network with our proposed model. The amount of photographs even has been increased by utilizing various image enhancement techniques. Ultimately, the transfer learn- ing methodology is utilized to boost the identification process accuracy even more. Five-fold cross-validating for SGD optimizer shows that the proposed model is capable of classifying the snake images with a high accuracy of 91.30%. Without Cross validation the model shows 90.50% accuracy. Keywords: Snake · CNN · Data augmentation · Deep learning · Transfer learning · Cross validation 1 Introduction Snakes are ectothermic, amniotic reptiles, surrounded in sepals, just like other squamates. Several snake species have skulls with a slew of joints than their reptile ancestral, allowing it to swallow predators with their extremely maneu- verable jaws relatively large unlike there own heads. There are two types of snakes such as non-venomous (non-poisonous snake) and venomous (poisonous N. Rezoana—equal contribution. c Springer Nature Switzerland AG 2021 M. Mahmud et al. (Eds.): AII 2021, CCIS 1435, pp. 216–231, 2021. https://doi.org/10.1007/978-3-030-82269-9_17
Snake Classification 217 snake). Venomous snakes are members of the suborder Serpents and are able to develop venom that they use to attack prey, defend themselves and help digest their prey. Utilizing hollow or grooved fangs, the venom is usually released by injection, while other venomous snakes lack well-developed fangs. Non-venomous snakes, except for massive constrictor snakes such as the Green Anaconda or the Burmese Python, are generally benign to humans. Like venomous snakes, non- venomous snakes have teeth. Snake envenoming is a major, worldwide common health issue with the greatest prevalence in Southeast Asia. Fig. 1. Top 5 countries with the highest rate of snake bite deaths per 100,000 people A analysis focused on 60 articles has reported that 363 victims with snake bites, both venomous (88%) and nonvenomous (12%) were diagnosed and treated if necrosis exists (15.2%) [30]. No infections were detected in patients although the antibiotics were not used. Thus, based on the analysis, it can be implied that antibiotics are present in the snake, considering the fact that very little raw data is given. Inability to identify snake from the visible characteristics is an important cause of mortality due to snake bite. For centuries, snake venom, particularly in Chinese culture, have also been used as medicine tools. Any of the leading drugs for high blood pressure, cardiovascular disease, and heart attacks used snake venom as a blueprint. As a consequence, snake venom is often known as a mini-drug repository, with each medicine being clinically effective. For example, the FDA has licensed medicines relying on snake venom, such as Captopril R (Enalapril), IntegrilinR (Eptifibatide), and Aggrastat R (Tirofiban) [24]. Aside from these approved medicines, various other snake venom materials for numerous thera- peutic applications are presently in pre-clinical or clinical trials. Snakebite envenoming is a well-known tropical disease exacerbated by a ven- omous snake accidentally injecting an extremely modified poison into humans. The snake’s fangs, that are remodeled teeth attached to a poison glands via a inlet are used to extract the potion. Nearly 2 million residents in Asia are
218 N. I. Progga et al. poisoned by snakes every year, although there are approximately 435,000 to 580,000 snake bites annually in Africa that require medication. Women, children and farmers in vulnerable remote regions in low and middle-income countries are affected by envenoming. The greatest risk exists in places where health care services are inadequate and medical facilities are insufficient. The detection of snake types is a key aspect of the diagnosis process. Classi- fication and detection of venomous and non-venomous snake has many dynamic applications such as: It can help the health professional in determining the best anti-venom to use. Snake venom is often known as a mini-drug repository, so by classifying the venomous snake from the non venomous snake the venom can be used to produce more drugs for mankind. 2 Related Work The Deep Convolution Neural Network (CNN) has carried out in previous decades a number of groundbreaking findings on image recognition, object iden- tification, semantic classification etc. [21]. Now it has become prominent in research for some day because it can handle an immense number of data [4]. CNN outperforms the machine learning approach in terms of efficiency. Classifi- cation and identification of images was an iconic machine learning issue, which was overcome through deep learning. In particular, technologies dealing with the visual information, including the biggest image classification data set (Image Net), computer vision and NLP analysis were quite impressive and the out- comes have been obtained is extraordinary. That is why we choose CNN for out study. The algorithms of deep learning are designed to imitate the role of the human brain [8]. In biological data [22,23] deep learning has been used. Also in object detection CNN has shown outstanding performance [11,27,37]. CNN has already shown remarkable effectiveness in diagnostic image recognition, such as the detection of lung cancer [14], skin cancer [28], oral cancer [18], brain tumors [10], and other diseases. By Using object detection and a machine learning method focused on Mask Region-based Convolutional Neural Network (Mask R-CNN) has been imple- mented for snake species classification [7]. To discern reptile species from images, deep learning, specifically CNN, has been used [5]. Snake sting marks images were also used to identify poisonous and non-poisonous snakes using the CNN method [20]. James et al. [16] presented a semiautomatic method in order to differentiate six distinct organisms by eliminating taxonomical characteristics from the pho- tos. There were 1,299 photographs in the dataset and 88 photos in the lowest frequent level. The outer edge taxonomical characteristics are least essential for organism recognition than the into front and side-view characteristics, according to various feature detection techniques. In object detection and image processing, various model architectures have been evaluated. To discern 5 specific snake varieties Abdurrazaq et al. [2] used 3 distinct Convolutional Neural Network (CNN) frameworks. In this study they
Snake Classification 219 worked with a set of 415 pictures. There were 72 photos required for the less common snake type. The maximum outcomes were achieved by a medium-sized classification framework. The Siamese network was used by Abeysinghe et al. [3] to categorize a com- paratively limited data set that contains 200 photographs of 84 organisms from the WHO. The method discussed in their paper focuses on one-shoot learning, as 3 to 16 photos per habitat were provided in the collection of data. The outcomes collected by the automatic categorization process is lower than the exactness of human identification accuracy. Snakebite poisoning is an overlooked environmental diseases which murders over 100,000 inhabitants and slaughters over 400,000 per year [36]. Snakebite is a frequent workplace hazard for citizens who make their living in cultivation, including those in South-East Communities in asia. An automated identifica- tion of a species of snake depending on its photograph has already been con- structed, as previously mentioned. As a classifier, a several supervised machine learning technique has been implemented like Naive Bayes, Decision Tree J48, k-Nearest Neighbors, or Back-Propagation Neural Network. The requirements in the features extraction process, on the other hand, are not easy to train in conventional machine learning techniques and must be manually calibrated. As a result, throughout this study, important contributions has made in the following areas: – A CNN based model has been proposed to classify venomous and non- venomous snake. – Various architectures have been compared in terms of performance. – K-fold cross validation has been applied on our proposed model with three different optimizer to improves generalization capacity. – The system can detect both venomous and non-venomous snakes in real time, allowing non-experts to recognize snake species with greater accuracy than previously mentioned approaches. 3 Data Pre-processing At first, we will discuss about the dataset. To initiate, the data set and model creation plan will be discussed in depth to aid in the planning of the proposed model. Following that, we’ll go through the proposed model architecture sim- ulation method in detail, as well as the training methodology for determining the best parameter modifications. Ultimately, we’ll use modeling approaches to demonstrate critical flaws in visual indicators in addition to creating a reported snake more identifiable. 3.1 About Dataset The set of data comes from kaggle.com and contains about 1766 snake images. Per photo was assigned to a category and was divided into groups by the respec- tive class labels such as non-venomous and venomous. Since reformatting is
220 N. I. Progga et al. among the most important phases of data preprocessing, all images are refor- matted to 224 × 224 pixels. Figure 3 and Fig. 4 depicts several photographs from the benchmark dataset. Figures 3 and 4 demonstrate that there are numerous differences between venomous and non-venomous snakes in terms of physical appearances such as head structure, eye shape, skin colour, and so on. The men- tioned features will aid our proposed model in learning the distinctions between poisonous and non-poisonous snakes. The set of data has been split into train, validation, and test segments in an appropriate proportion Fig. 2. Non-venomous snake image Fig. 3. Venomous snake image 3.2 Data Augmentation It is well established that even a massive quantity of data in the datasets is needed to achieve a better result for a CNN model. The data augmentation techniques is necessary for correctly implementing a CNN architecture. This methodology prevents data manipulation and maintains the initial reliability. This technique is often used during the training process to increase the efficiency of the architecture by fixing overfitting problems. If the dataset is large enough, several features can be extracted from it and compared to unidentified data. However, if there is insufficient data, data augmentation may be implemented to boost the model’s accuracy [9,15,26,33]. Through applying augmentation oper- ations to training images, such as random rotation, shift, zoom, noise, flips, and so on, data augmentation will generate multiple pictures [33]. Every parame- ter has the ability to represent photos in a number of aspects and come up
Snake Classification 221 with particular features during the training phase, increasing the framework’s effectiveness. Since our dataset is smaller in size, we implemented a variety of augmentation functions on it. For the augmentation, Image Data Generator was utilized. Figure 5 represents the primary picture as well as the augmented photos created from that. We implemented the model with 80% of the overall of the pictures and used the remaining 20% to validate the system throughout evalua- tion. The settings for image augmentation used in our experiment can be seen in Table 1. Fig. 4. Data augmentation Table 1. Images augmentation settings Augmentation setting Range Rotation 0.2 Zoom 0.1 Contast 0.1 Horizontal flip T rue 4 Methodology Convolutional neural networks is inspired by neurological mechanisms. A convo- lutional neural system is made up of many layers, including convolution layers, pooling layers, and fully connected layers, and it uses a back-propagation algo- rithm to obtain features to train the model properly. Figure 6 depicts the overall research’s system flow chart. 4.1 Model Construction In this study, the framework was implemented using a Convolutional Neural Net- work (CNN) and data augmentation. First, the architecture uses the dataset to take the pictures. Then preprocessing begins. Then some augmentation param- eters are used to enlarge the dataset. Ultimately, the enlarged set of data is
222 N. I. Progga et al. Fig. 5. System flowchart used to forecast the class by the CNN architecture. The pictures were stan- dardised to some extent to be properly categorized. CNN itself performed the characteristic retrieval of the pictures. We’ll describe our recommended model architecture in detail in this section. The proposed method includes three basic components: feature extraction, identification, and classification. In the begin- ning, we include synchronized layers of convolution, activation, and max-pooling in our system design. The features are then passed into a flattened layer. The flattened attributes or features are then transferred into dense layers. In the dense layer dropout has been used to avoid overfitting. The classification pro- cess was then completed by the ultimate layer, which stated the softmax layer. The photos were provided to the model for the training process after the data augmentation. There is a CBr = iconv block in the system built, accompanied via a max pooling stratum that is three times in consonance in sequential format. Maxpooling is an efficient way of downsizing the tightly-scale photos, requiring maximal rates for each stratum, since most of the features generated are ignored by utilizing a 3×3 filter size. As discussed during the preceding analysis superim- posed max pooling frames do not significantly increase over the nonoverlapping windows, so the max pooling layers used during our experiment was 2 × 2 with stride 2. It consists of 3 convolutionary layers, with 16 filters of 3 × 3 in the first convolutional layer,32 filters in the second convolutional layers and the third layer is 64 layers 3 × 3 in dimension. The scale of the kernel is 2. The produced features are reused by the attached activation function in the early stages to construct an unique feature map as output just after the convolution layer. In equation, convolution over an image q(x, y) is defined using a filter p(x, y). j k p(x, y) ∗ q(x, y) = w(m, n)f (x − m, y − n) (1) y=−j y=−k To generate the low-level features all layers of convolution was accompanied by the ReLU activation function. Following the previous convolution layers, the ReLU activation function was used in the hidden layer as well. ReLU has a number of advantages, the most notable of which is its ability to quickly dis- tribute gradients. As a result, calculating the essential characteristics of CNN in the provisional mass ,reduces the chances of gradient extinction. The activat- ing mechanism tends to perform component-by-component operations on this given input feature map, since the result tends to be the identical dimension as
Snake Classification 223 Fig. 6. Model architecture the origin. As a result, among the most common activation functions, ReLU is being used in the all layers. ReLU [12] has been included to demonstrate that the architecture does not experience linearity, as shown in the corresponding formula: R(x) = max(0, x) (2) The architecture is then exposed in the flattened layers to convert the feature map. From feature map generated by the previous convolution layers to complete the categorization task, a single dimensional feature vector has been constructed by flattened layer. A drop-out layer is linked following the Fully connected or hidden layer and the size of the fully connected layer is 128. This dropping layer will dump the weights of the Fully connected layer at random during training to mitigate unnecessary weight and overfitting. For our CNN model, we choose 0.2 as the dropout range for the drop out layer. The dropout layer’s job is to discard 20% of the nodes in each FC layer in order to prevent overfitting [29]. The hyperparameters were fine-tuned by layering again until fault did not radically alter. In addition, the range of the dropout for the proposed model was determined by experimenting for several constants. The drop - outs range of 0.2 was considered as the best amount of the dropout layer because it tended to prevent overfitting rather than others value. In order to maximize the outcome the parameters are being adapted. Ultimately, since there are two categories the output layer contains two nodes. The softmax [32] activation function was used just after the FC layer, as shown in the following equation, to classify snake images into non-venomous and venomous categories. Figure 8 portrays the representations of the suggested model’s layers.
224 N. I. Progga et al. ei Sof tmax(x) = i (3) i=0 ei Fig. 7. The proposed model’s detailed layer representation 4.2 The Implementation Procedure The coding for the framework was produced and implemented in Google Colab [6] using the python programming language. Keras [13], Tensorflow [1], NumPy [35], and Matplotlib [31] were the libraries utilized in the whole study. The back- end of the framework was chosen as Tensorflow, and keras has been utilized to offer additional built-in functionality such as activation functions, optimizers, layers, and so on. Keras API was used to enhance the dataset. NumPy is a Python library for mathematical evaluation. Confusion matrix, split train and test files, modelcheckpoint, callback mechanism, as well as other schematic rep- resentation like confusion matrix, loss against epochs graphs, accuracy against epochs curves, and many more, are all generated using Sklearn. The matplotlib library is also needed to create visual representations of the previously mentioned diagrams, such as the confusion matrix. 5 Experimental Evaluation The performance of the implemented model to classify photos of snakes specif- ically grouped into two groups, non-venomous and venomous, will be covered in this section. In addition, we will compare our proposed model to other tra- ditional models such as Inception Net, Resnet50, VGG 16, VGG 19, Xception, MobileNet v2 and Inception Resnet V2. The outcomes of our model’s k fold cross validation for different optimizers will also be articulated.
Snake Classification 225 5.1 Tuning of the Hyper-parameters Since fine-tuned hyper-parameters also have major impact on the CNN archi- tecture’s performance and they are are essential because they strongly impact the model’s attitude. The Adam, Adamax, and SGD optimizers were utilized to trainIing 100 epochs for the proposed model, with a learning rate of 0.0001 with the batch size of 32. K fold cross validation has also been performed for the various optimizers. As the loss function categorical cross entropy was being used, the loss of class probability caused by the softmax function was also determined by this loss function. Finally, calculate each category’s probability of occurrence. ModelCheckpoint has been added as a callback function as well. 5.2 K-Fold Cross Validation To test our model, we attempt 5 fold cross-validation. Due to the general opera- tional overhead, this phase is normally avoided in CNNs. The dataset is divided into three sections using K-folds cross-validation. Fold1 is composed of part 1 as a training set, part 2 as a validation set, and part 3 as testing set, while fold2 is consists of part 2, part 1 and part 3 as a training, validation and testing set respectively. Thus, the fold continues until it reaches 5 folds, with each fold containing unique training, validation, and testing datasets. The K-fold cross validation approach accounts for the utilization of various training and test- ing data, which reduces overfitting and improves generalization capacity. As a consequence, we may generalize our outcomes over the dataset. 5.3 Result Figure 8 depicts the accuracy of the proposed model as compared to other stan- dard Convolution neural networks like Inception Net, VGG 19, Resnet50, Xcep- tion Net, MobileNet v2, Inception Resnet V2 and VGG 16. Despite providing a dataset with a limited number of photographs, the designed model produces a high categorization accuracy cpmpared to other CNN models. As shown in the Fig. 9 Inception Net, VGG 19, Resnet50, Xception Net, MobileNet v2, Inception Resnet V2 and VGG 16 has 82.38%, 43.75%, 81.81%, 80.94%, 82.35%, 89.62% and 62.50% accuracy respectively. Our model outperformed then these models with 90.50% accuracy. The best result corresponds to the model we introduced, based on the current Inception Net, VGG 19, Resnet50, Xception Net, MobileNet v2, Inception Resnet V2, and VGG 16 models. As a result, it can be concluded that the developed framework outperforms the other ones. The performance of all evaluation models, as seen in Fig. 8, indicate that the method is superior to others. Figure 9 compares the proposed model’s accuracy and loss curve to those of other traditional CNN models. From the shape of the curves in this mentioned figure, we can see that other convolutional models have a propensity to overfit or underfit, while our proposed model seems to have a good fit. Our proposed model’s high accuracy and low loss have already been considered the best case scenario for any CNN model. Figure 10 shows the detection performance of the
226 N. I. Progga et al. Fig. 8. Accuracy of proposed model & other traditional models proposed model and shows the detection result of the test dataset’s in two classes: venomous and nonvenomous. In comparison to the previously listed other CNN models, 40 images were arbitrarily checked and 9 of those being seen in Fig. 6. The identification performance for the developed method is really exceedingly decent. The confusion matrices of the Inception Net, VGG 19, Resnet50, Xcep- tion Net, MobileNet v2, Inception Resnet V2, and VGG 16 models, as well as the suggested system, are shown in Fig. 11. For our adapted CNN model, the diagonal magnitude of the confusion matrices in both classes is greater than the other models. That is, the proposed model will precisely distinguish the same number of test samples from our test dataset as the current Neural network model. As a result, our model successfully outperformed the other conventional Neural network model in this area as well. Finally, Fig. 12 demonstrates the K-fold cross validation result of the proposed model utilizing three optimizers: Adamax, Adam, and SGD. By using adamax optimizer for K fold cross vali- dation the validation accuracy and the testing accuracy for fold-1 is 83.58% & 86.87%, for fold-2 it is 81.83% & 82.52%, for fold-3 it is 77.44% & 80.07%, for fold-4 it is 84.24% & 84.96& and for fold-5 it is 86.87% & 83.20% respectively. For adam optimizer the validation accuracy and the testing accuracy for fold-1 is 86.56% & 89.22%, for fold-2 it is 87.57% & 88.04%, for fold-3 it is 83.35% & 84.42%, for fold-4 it is 87.78% & 90.48% and for fold-5 it is 84.36% & 88.43% respectively. For SGD optimizer the validation accuracy and the testing accu- racy for fold-1 is 84.85% & 85.78%, for fold-2 it is 88.30% & 89.67%, for fold-3 it is 84.78% & 85.61%, for fold-4 it is 83.77% & 88.04% , and for fold-5 it is 87.73% & 91.30 respectively. The SGD optimizer in fold-5 produces better outcomes for the implemented model, as well as the other results are also quite satisfactory. Also it demonstrates that the accuracy of the classification model is unaffected by the training data.
Snake Classification 227 (a) InceptionNet (b) VGG19 (c) Resnet50 (d) Xception (e) Mobilenet v2 (f) Inception- (g) VGG16 (h) Proposed Resnet-v2 Model Fig. 9. Accuracy and loss curve Fig. 10. Result of detection produced by the proposed model
228 N. I. Progga et al. (a) InceptionNet (b) VGG19 (c) Resnet50 (d) Xception (e) Mobilenet v2 (f) Inception (g) VGG16 (h) Proposed Resnet v2 Model Fig. 11. Confusion matrix of the models (a) adamax Optimizer (b) Adam Optimizer (c) SGD Optimizer Fig. 12. K-Fold cross validation for proposed model using different optimizer 6 Epilogue and Future Work A new convolutional neural network-based architecture for detecting and classi- fying venomous and non-venomous snakes was suggested during this whole study. The framework’s ability to acquire snake features using neural network blocks is clearly demonstrated. In comparison to the various possibly the best Convolu- tional neural network frameworks Inception Net, VGG 19, Resnet50, Xception Net, MobileNet v2, Inception Resnet V2, and VGG 16, the architecture has remarkable categorization accuracy. This study looks at how to develop and create a venomous and nonvenomous snake classification model that could help mankind. Snake venomous can be used as medicinal tools by distinguishing between venomous and non-venomous snakes. Snakebite disease may be minimized by identifying the species of snake and administering appropriate treatment.
Snake Classification 229 The suggested solution greatly outshines state-of-the-art frameworks, with a dramatic increase in accuracy of 90.50%, according to the experimental research review. In addition, the model performs admirably in terms of K fold cross val- idation outcomes. It would have been more useful in upcoming analysis if we perform a grid hunt on the hyper-parameters and find the most suitable num- ber of parameter values for K fold cross validation. Finally, we believe that the current findings would hopefully resolve unique stimulation in recognizing addi- tional snake images and categorizing them in Artificial intelligence based envi- ronments, particularly in medical application. To create our suggested structure more reliable and authenticated, more analysis should be done to evaluate and gather a diverse data set with significant quantities of information about snake photographs. Furthermore, many additional Neural networks, such as DenseNet, EfficientNet, and others, can be used to enhance information horizons, and the model’s output can be compared to that of other mainstream machine learning techniques. We will try to include more pictures of snake in future captured in different environment and lightening to analyze the model performance on those images. Also an integration of data-driven (CNN) and knowledge-driven (BRBES) approach can be proposed to analyze risk assessment of a snake bite in the human body [17,19,25,34]. References 1. Abadi, M., et al.: Tensorflow: a system for large-scale machine learning. In: 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 2016), pp. 265–283 (2016) 2. Abdurrazaq, I.S., Suyanto, S., Utama, D.Q.: Image-based classification of snake species using convolutional neural network. In: 2019 International Seminar on Research of Information Technology and Intelligent Systems (ISRITI), pp. 97–102. IEEE (2019) 3. Abeysinghe, C., Welivita, A., Perera, I.: Snake image classification using siamese networks. In: Proceedings of the 2019 3rd International Conference on Graphics and Signal Processing, pp. 8–12 (2019) 4. Albawi, S., Mohammed, T.A., Al-Zawi, S.: Understanding of a convolutional neu- ral network. In: 2017 International Conference on Engineering and Technology (ICET), pp. 1–6. IEEE (2017) 5. Annesa, O.D., Kartiko, C., Prasetiadi, A., et al.: Identification of reptile species using convolutional neural networks (CNN). Jurnal RESTI (Rekayasa Sistem Dan Teknologi Informasi) 4(5), 899–906 (2020) 6. Bisong, E.: Google colaboratory. In: Building Machine Learning and Deep Learning Models on Google Cloud Platform, pp. 59–64. Springer (2019) 7. Bloch, L., et al.: Combination of image and location information for snake species identification using object detection and efficientnets. CLEF working notes (2020) 8. Chauhan, R., Ghanshala, K.K., Joshi, R.: Convolutional neural network (CNN) for image detection and recognition. In: 2018 First International Conference on Secure Cyber Computing and Communication (ICSCCC), pp. 278–282. IEEE (2018)
230 N. I. Progga et al. 9. Chowdhury, R.R., Hossain, M.S., Ul Islam, R., Andersson, K., Hossain, S.: Bangla handwritten character recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th International Conference on Informatics, Elec- tronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 318–323. IEEE (2019) 10. Dı́az-Pernas, F.J., Martı́nez-Zarzuela, M., Antón-Rodrı́guez, M., González-Ortega, D.: A deep learning approach for brain tumor classification and segmentation using a multiscale convolutional neural network. In: Healthcare, vol. 9, p. 153. Multidis- ciplinary Digital Publishing Institute (2021) 11. Galvez, R.L., Bandala, A.A., Dadios, E.P., Vicerra, R.R.P., Maningo, J.M.Z.: Object detection using convolutional neural networks. In: TENCON 2018–2018 IEEE Region 10 Conference, pp. 2023–2027. IEEE (2018) 12. Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: Pro- ceedings of the fourteenth international conference on artificial intelligence and statistics, pp. 315–323. JMLR Workshop and Conference Proceedings (2011) 13. Gulli, A., Pal, S.: Deep Learning with Keras. Packt Publishing Ltd., Birmingham (2017) 14. Heuvelmans, M.A., et al.: Lung cancer prediction by deep learning to identify benign lung nodules. Lung Cancer 154, 1–4 (2021) 15. Islam, M.Z., Hossain, M.S., Ul Islam, R., Andersson, K.: Static hand gesture recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 324–329. IEEE (2019) 16. James, A.P., Mathews, B., Sugathan, S., Raveendran, D.K.: Discriminative his- togram taxonomy features for snake species identification. HCIS 4(1), 1–11 (2014) 17. Jamil, M.N., Hossain, M.S., Ul Islam, R., Andersson, K.: A belief rule based expert system for evaluating technological innovation capability of high-tech firms under uncertainty. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 330–335. IEEE (2019) 18. Jeyaraj, P.R., Nadar, E.R.S.: Computer-assisted medical image classification for early diagnosis of oral cancer employing deep learning algorithm. J. Cancer Res. Clin. Oncol. 145(4), 829–837 (2019) 19. Kabir, S., Islam, R.U., Hossain, M.S., Andersson, K.: An integrated approach of belief rule base and deep learning to predict air pollution. Sensors 20(7), 1956 (2020) 20. Kamalraj, R.: Deep learning model for identifying snakes by using snakes’ bite marks. In: 2020 International Conference on Computer Communication and Infor- matics (ICCCI), pp. 1–4. IEEE (2020) 21. Li, Y., Hao, Z., Lei, H.: Survey of convolutional neural network. J. Comput. Appl. 36(9), 2508–2515 (2016) 22. Mahmud, M., Kaiser, M.S., McGinnity, T.M., Hussain, A.: Deep learning in mining biological data. Cogn. Comput. 13(1), 1–33 (2021) 23. Mahmud, M., Kaiser, M.S., Hussain, A., Vassanelli, S.: Applications of deep learn- ing and reinforcement learning to biological data. IEEE Trans. Neural Netw. Learn. Syst. 29(6), 2063–2079 (2018) 24. Mohamed Abd El-Aziz, T., Soares, A.G., Stockand, J.D.: Snake venoms in drug discovery: valuable therapeutic tools for life saving. Toxins 11(10), 564 (2019)
Snake Classification 231 25. Monrat, A.A., Islam, R.U., Hossain, M.S., Andersson, K.: A belief rule based flood risk assessment expert system using real time sensor data streaming. In: 2018 IEEE 43rd Conference on Local Computer Networks Workshops (LCN Workshops), pp. 38–45. IEEE (2018) 26. Progga, N.I., Hossain, M.S., Andersson, K.: A deep transfer learning approach to diagnose covid-19 using x-ray images. In: 2020 IEEE International Women in Engi- neering (WIE) Conference on Electrical and Computer Engineering (WIECON- ECE), pp. 177–182. IEEE (2020) 27. Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detec- tion with region proposal networks. arXiv preprint arXiv:1506.01497 (2015) 28. Rezaoana, N., Hossain, M.S., Andersson, K.: Detection and classification of skin cancer by using a parallel CNN model. In: 2020 IEEE International Women in Engi- neering (WIE) Conference on Electrical and Computer Engineering (WIECON- ECE), pp. 1–7. IEEE (2020) 29. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014) 30. Terry, P., Mackway-Jones, K.: Antibiotics in non-venomous snakebite. Emerg. Med. J. 19(2), 142–142 (2002). https://emj.bmj.com/content/19/2/142.1 31. Tosi, S.: Matplotlib for Python Developers. Packt Publishing Ltd., Birmingham (2009) 32. Tüske, Z., Tahir, M.A., Schlüter, R., Ney, H.: Integrating gaussian mixtures into deep neural networks: softmax layer with hidden variables. In: 2015 IEEE Inter- national Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4285–4289. IEEE (2015) 33. Uddin Ahmed, T., Hossain, M.S., Alam, M., Andersson, K., et al.: An integrated CNN-RNN framework to assess road crack. In: 2019 22nd International Conference on Computer and Information Technology (ICCIT) (2019) 34. Ul Islam, R., Andersson, K., Hossain, M.S.: A web based belief rule based expert system to predict flood. In: Proceedings of the 17th International conference on information integration and web-based applications & services, pp. 1–8 (2015) 35. Van Der Walt, S., Colbert, S.C., Varoquaux, G.: The numpy array: a structure for efficient numerical computation. Comput. Sci. Eng. 13(2), 22–30 (2011) 36. Warrell, D.A., et al.: Guidelines for the management of snake-bites. Guidelines for the management of snake-bites (2010) 37. Yanagisawa, H., Yamashita, T., Watanabe, H.: A study on object detection method from manga images using CNN. In: 2018 International Workshop on Advanced Image Technology (IWAIT), pp. 1–4. IEEE (2018)
You can also read