Aim This research work aims at developing an automatic medical image analysis and detection for accurate classification of brain tumors from a magnetic resonance imaging (MRI) dataset. We developed a new MIDNet18 CNN architecture in comparison with the AlexNet CNN architecture for classifying normal brain images from brain tumor images.
Materials and methods The novel MIDNet18 CNN architecture comprises 14 convolutional layers, seven pooling layers, four dense layers, and one classification layer. The dataset used for this study has two classes: normal brain MR images and brain tumor MR images. This binary MRI brain dataset consists of 2918 images as the training set, 1458 images as the validation set, and 212 images as the test set. The independent sample size calculated was seven for each group, keeping GPower at 80%.
Result From the experimental performance metrics, it could be inferred that our novel MIDNet18 achieved higher test accuracy, AUC, F1 score, precision, and recall over the AlexNet algorithm.
Conclusion From the result, it can be concluded that MIDNet18 is significantly more accurate (independent sample t-test P<0.05) than AlexNet in classifying tumors from brain MRI images.
Key words: AlexNet, binary classification, brain tumor image, convolutional neural network, deep learning, novel medical image analysis and detection network
*Corresponding author: Ramya Mohan, Saveetha School of Engineering, Saveetha Institute of Medical and Technical Sciences, Saveetha University, Chennai, India. Email: ramyanallu@gmail.com
Submitted: 19 December 2021; Accepted: 28 January 2022; Published: 18 March 2022
©2022 Mohan R, et al.
This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0). License (http://creativecommons.org/licenses/by-nc/4.0/)
Brain tumor disease5,11,18 is increasing day by day due to various factors. Nowadays, upon early diagnosis, the usage of modern tools and technology helps to increase the survival rate/period of brain tumor patients. Age groups between 20 and 39 years have a high percentage of survival gain. Few brain tumor categories show less survival rate in young and older age groups.3–5,18 Poor diagnosis and classification of the tumor also decreases the survival rate. Early diagnosis and proper severity classification help to reduce the mortality rate of a large population8 to an extent.
Artificial intelligence (AI) in machines has human-like intelligence to learn, solve, and reason any disease when presented with an enormous variety of data.4,27 Of late, the impact of AI in diagnosis and treatment of brain tumor-related diseases has been increasing drastically. AI is used for brain tumor surgery, pediatric brain tumor imaging,9 brain tumor segmentation,21 and differentiating brain tumor from non-neoplastic lesions.1 Therefore, AI has been adopted for brain tumor classification for effective treatment and early diagnosis.
Kesav and Jibukumar (2021) have used two channel, CNN and RCNN, models to classify glioma and normal brain magnetic resonance images.13 Also, this work detected meningioma and pituitary tumors with high average confidence levels. Tandel et al. (2021) used brain tumor grading methods with deep learning and machine learning techniques.23 This work also compared the same with five different CNN models. Özcan et al. (2021) introduced customized CNN for grading the degree of malignancy in brain tumor.20 Clinical cases and augmented dataset were used for training the data. The work was effective and robust in classification of low-grade and high-grade gliomas. Khan et al. (2021)14 performed brain tumor segmentation using the K-means clustering algorithm and classification of MR images using the VGG19 CNN model. Khan et al. (2021)14 and Kang et al. (2021)12 applied transfer learning-based deep CNN to extract features from brain MR images. A support vector machine with radial basis function outperforms the CNN model. Ismael (2018)10 designed a CADD system to classify abnormal growth in brain images. A set of features are obtained from discrete wavelet transform and Gabor filter method. Stacked sparse autoencoder and softmax classifier is used for identifying three brain tumors. Ismael (2018)10 and Lotlikar et al. (2021)21 presented a review work that presents preprocessing techniques, deep learning techniques and machine learning algorithms used by researchers with challenges encountered in performing analysis.
The study was conducted in the AI research laboratory in Saveetha School of Engineering, Saveetha Institute of Medical and Technical Sciences. The dataset was downloaded from Kaggle (“Kaggle: Your Machine Learning and Data Science Community” n.d.). Since it was downloaded from the public database, no ethical clearance was necessary. There are two groups involved in this study. Based on,21 the sample size calculated for the study was 14 with parameters Alpha 0.05, beta 0.2, and g-power 0.8 (Figure 1).
Figure 1. Proposed MIDNet18 architecture.
The dataset consists of 2918 images belonging to two classes (presence of tumor and normal brain MRI) under the training folder. The validation folder consists of 1458 augmented images belonging to both classes. Further, 212 separate images belonging to both classes were kept in the test folder, which was duly marked by medical experts from the Saveetha Medical College and Hospital. This study was performed on a novel medical image analysis and detection network (MIDNet18) CNN architecture, and its results were compared with AlexNet architecture.
The study was conducted on a MacBook Air with an Apple M1 chip and 8 GB memory. All of the CNN models were run in Google Colab, which provides a single 12 GB NVIDIA Tesla K80 GPU. All analyses are conducted using SPSS software.15 Independent variables in this study are the input variables (brain tumor and nontumor MRI images. The dependent variables are output variables (accuracy, precision, recall, F1 score). Independent t-test is performed to compare the performance of algorithms.
Figure 1 shows the architecture of our proposed MIDNet18 model. The MIDNet18 model consists of 14 convolutional layers with a 3×3 kernel size. The ReLu activation function was used in all of the layers. These convolutional layers provide a feature map from the image. The input size of the image is maintained as 224×224. The model consists of seven max pooling layers with a pooling size of 2×2. Max pooling19 in the MIDNet18 CNN model helps in highlighting the brighter pixels.7 Batch normalization is performed in the MIDNet18 model, which helps in avoiding the overfitting of the model. In simple terms, batch normalization helps each layer to learn more independently. Four Dense layers with
Figure 2. AlexNet architecture.
The first convolutional network using GPU to boost performance was AlexNet. It consists of five convolutional layers, three max pooling layers, two normalization layers, two fully connected layers, and one softmax layer. The convolutional layers contain two convolutional filters and a ReLu activation function. The input size is generally 224×224×3, but due to the occurrence of padding, it is actually 227×227×3.
Begin
Step 1: Dataset consisting of normal and brain tumor images of size 224×224 is uploaded to the MIDNet18 model as input. Total data is split as training dataset for learning the images, validation dataset to validate the trained brain images, and test dataset which is data that is not learned by the model.
Step 2: The MIDNet18 input layer consisting of the training dataset is given to the convolutional layer of the CNN model.
Step 3: Fourteen convolutional layers are implemented in the MIDNet18 model. Each convolution layer is convoluted with a kernel size of 3×3 with one stride. Default padding is used in each layer. The convolutional layer uses ReLu activation function in order to obtain feature maps.
Step 4: Batch normalization is carried out for each convolutional layer in order to range the value between 0 and 1.
Step 5: The output after convolution is given to max pooling for selecting the best or the max value.
Step 6: Max pooling of all 14 convolutional layers gives feature maps as input to the dense layer.
Step 7: The dense layer has three hidden layers and one output layer. The hidden layers in a fully connected network learn the brain images by updating the weights in 100 iterations.
Step 8: The output layer classifies the trained data as a normal or tumor brain image.
Step 9: The process is validated and tested with validation and test dataset for understanding the model performance.
End
The training and validation loss of MIDNet18 is represented in Figure 3. It could be inferred from the figure that the loss was as high as 90% before the 15th iteration. But as the iteration increases, the learning capacity of the model with its weight updation algorithm reduces the loss parameter. This reduction in the training and the validation was maintained until the 100th epoch.
Figure 3. Representation of the training loss and validation loss performance of MIDNet18 in different iterations.
Figure 4 represents the training and the validation loss of AlexNet. It is observed that the initial training loss in AlexNet was more than 70% and the validation loss was 60%. Around the 10th iteration, it could be noticed that there was a steep fall in the loss reaching around 15%. With the increase in the number of iterations, the loss percentage of both the training and validation was below 10% and was maintained till the 100th iteration.
Figure 4. Representation of the training loss and validation loss performance of AlexNet in different iterations.
Training and validation accuracy performance of MIDNet18 is represented in Figure 5. It can be observed that initially the training and validation accuracy of MIDNet18 was as low as 50% and 70%, respectively. But the training accuracy increased to 90% around the third or fourth iteration, and from the graph, it can be noticed that the training accuracy was maintained at around 99% until the 100th iteration. With respect to the validation accuracy, until 20 iterations, the accuracy shifted from 85% to 65% and then again increased to 96%, followed by a slight slip to around 88%. But from the 20th iteration, the accuracy of the validation dataset was maintained at around 99 for the remaining epochs.
Figure 5. Representation of the training and validation accuracy performance of MIDNet18 in different iterations.
From Figure 6, it can be inferred that the training accuracy of AlexNet was 70% initially, which gradually increased to more than 90% at around the 18th iteration. AlexNet maintained the training accuracy of more than 98% for the remaining 82 epochs. The same pattern could also be seen for the validation accuracy, which also started with as low as 70% accuracy. But around the 20th iteration, it spiked above 90% and thereafter gradually built up the accuracy to more than 90%.
Figure 6. Representation of the training and validation accuracy performance of AlexNet in different iterations.
It can be observed from Figure 7, that the training AUC in MIDNet18 reaches 99% around the 5th iteration and remains constant throughout the entire iterations. Similarly, the validation AUC of MIDNet18 also increased after the 5th iteration, but there was a sudden drop in AUC around the 50th iteration, which was regained immediately to 99% in the following epoch due to the MIDNet architecture’s resilient nature. For the rest of the epochs, the 99% AUC was maintained by the novel MIDNet model.
Figure 7. Training and validation area under curve (AUC) of the MIDNet18 model.
The training and validation AUC of AlexNet are represented in Figure 8. From the figure, it can be noticed that training and validation AUC of AlexNet started around 75% and 80%, respectively. As the number of epochs increased, both the training and validation AUC went up to around 97.5% and was maintained for the remainder of the 100 epochs.
Figure 8. Training and validation AUC (area under curve) of the AlexNet model.
Figure 9 represents the training and validation F1 score of the MIDNet18 model. From the figure, it can be inferred from the training and validation dataset that the F1 score reaches as high as 100% in the first few iterations itself. But around the 50th iteration, both the training and the validation F1 score became as low as 20% and 70%, respectively, which could also be noticed in the drop of AUC (Figure 7) around the same iteration. This sudden decrease was overcome by the model very shortly, and it regained close to 100%, maintaining it for the rest of the iterations. The training and validation F1_score of AlexNet (Figure 10) shows that the model reaches as high as 94.25% during the 20th iteration and maintains it stably thereafter till the last iteration.
Figure 9. Training and validation: F1 score of the MIDNet18 model.
Figure 10. Training and validation F1 score of AlexNet.
The training and precision of MIDNet18 are represented in Figure 11. It can be inferred that the precision of MIDNet18 for training and validation dataset reached nearly 99% during the first few iterations. The precision of the validation dataset alone dropped at around the 50th iteration, but due to the batch normalization and learning rates of MIDNet18, the model was able to perform the weight adjustment appropriately, which not only helped in regaining the percentage back to 99% but also in maintaining it until the 100th iteration. This was well reflected in the test dataset where the MIDNet model achieved the precision score of 98.78% (Table 1) when compared with that of AlexNet model’s 93.90%. It can be observed from Figure 12 that for both the training and validation, the AlexNet model established the precision score close to 100%, which remained constant with an increase in the number of iterations. But the model was able to achieve the precision score of 93.90% (Table 1) for the test dataset.
Table 1. Comparison of MIDNet18 and AlexNet CNN on Various Metrics, Namely, Training Accuracy, Testing Accuracy, Training Loss, Testing Loss, AUC, F1 Score, Precision, and Recall Value, for Classification of Tumor from Brain MRI Images.
Architectures | Training Acc (%) | Testing Acc (%) | Testing Loss (%) | AUC (%) | F1 Score (%) | Precision (%) | Recall (%) |
---|---|---|---|---|---|---|---|
Our Novel MIDNet-18 | 99.42 | 98.78 | 02.01 | 99.98 | 98.79 | 98.78 | 98.78 |
AlexNet | 99.07 | 93.90 | 50.34 | 95.41 | 94.25 | 93.90 | 93.90 |
Figure 11. Training and validation: precision of MIDNet18.
Figure 12. Training and validation precision performance metric of AlexNet.
Figure 13 represents the recall of the MIDNet18 CNN model. It could be inferred from the figure that the recall of the model is similar to the precision. The MIDNet model maintained around 99% for the training and validation. During validation, the recall in MIDNet18 showed a sudden decrease around the 50th iteration but increased to 98% shortly after that. The model achieved a test recall of 98.78%, which was 5% higher than that of the AlexNet model (Table 1).
Figure 13. Training and validation recall of MIDNet18.
A comparison of MIDNet18 and AlexNet for all the performance metrics is given in Table 1. The proposed MIDNet18 model achieves a higher testing accuracy of 98.78% in comparison with AlexNet’s testing accuracy of 93.90%. In comparison with the loss performance metrics, MIDNet18 achieved a low testing loss of 2.01 when compared with AlexNet’s testing loss of 50.34%. Similarly, the area under curve (AUC) for AlexNet is 95.41%, which is 4.57% lower than that of the MIDNet18 model. MIDNet18 achieved the F1 score of 98.79% in comparison with AlexNet, which displayed only 94.25%. MIDNet18 achieved higher precision and recall with a difference of 5% when compared with AlexNet. All of the performance metrics indicate that the accuracy of the MIDNet18 CNN model is better in classifying brain tumor images from the brain MRI dataset compared to the AlexNet model.
The training and the validation recall of AlexNet (Figure 14) indicate that the model maintained around 98% throughout the iterations. It could be inferred from Table 1 that the model achieved a recall of 93.90% for the test dataset (Table 1), which was lesser than that of the MIDNet18 model.
Figure 14. Training and validation recall performance metrics of AlexNet.
The bar chart (Figure 15) compares the mean accuracy of AlexNet and MIDNet convolutional neural network models for classifying tumor lesions from the brain MRI dataset.
Figure 15. Bar chart comparing the mean accuracy of AlexNet and MIDNet convolutional neural network models shows that MIDNet18 is significantly more accurate than AlexNet in classifying tumors from the brain MRI images.
We observe that MIDNet18 is significantly more accurate than AlexNet in classifying the tumors from brain MRI images, which can also be observed from the statistical test conducted (independent sample t-test P<0.05) indicating that the MIDNet18 model prediction accuracy is statistically significant, as shown in Tables 2 and 3.
Table 2. Comparison of the Accuracy of MIDNet versus AlexNet Algorithms in Detecting Cancer Lesions in Brain MRI Images (Independent Sample t-Test P < 0.05).
Independent Sample Test | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
Levene’s Test For Equality of Variances | t-Test for Equality of Means | |||||||||
F | Sig. | t | df | Sig. (2-tailed) | Mean Difference | Std. Error Difference | 95% Confidence Interval of the Difference | |||
Lower | Upper | |||||||||
Accuracy | Equal variances assumed | 6.283 | .013 | –1.823 | 198 | .070 | –1.73310 | .95087 | –3.60824 | .14204 |
Equal variances not assumed | –1.823 | 172.345 | .070 | –1.73310 | .95087 | –3.60996 | .14376 |
MIDNet18 superior accuracy in prediction over AlexNet is statistically significant.
Table 3. Statistical Comparison of AlexNet and MIDNet18.
Group statistics | |||||
---|---|---|---|---|---|
Algorithm | N | Mean | Std. Deviation | Std. Error Mean | |
Accuracy | AlexNet | 100 | 95.1532 | 7.91519 | .79152 |
MidNet18 | 100 | 96.8863 | 5.26932 | .52693 |
MIDNet18 Gives the Highest Mean Accuracy of 96.88% over AlexNet.
Liu et al. (2022)16 distinguished benign inverted papilloma tumors using a three-dimensional CNN model. All-Net achieved accuracy of 77.9% and sensitivity and specificity of 66.7% and 81.5%, respectively. Wang et al. (2021)26 identified nodal metastasis stage for lung cancer prediction using a multi-energy-level fusion model with principal feature enhancement. This fusion model achieved an accuracy of 93% and the kappa value of 86%. Arshad et al. (2021) performed skin cancer classification using ResNet-50 and ResNet-101. The standard models used achieved an accuracy of 91.7%. Gab Allah, Sarhan, and Elshennawy (2021) performed classification of brain tumors using VGG19 coupled with classifiers. This framework gave an accuracy of 98.54% with high computational time. Tazin et al. (2021) identified brain tumor with X-ray images using CNN and transfer learning techniques and achieved an accuracy of 92% in MobileNetV2, 91% in inceptionV3, and 88% in VGG19. Saba et al. (2021) used a DenseNet201-based CNN model for breast tumor diagnosis and obtained an accuracy of 92.8%. Saba et al. (2021)22 and Wang et al. (2021)25 used the ResNet34 model for classification of benign and nonbenign tumors. The overall accuracy obtained using this model was 75%. In observing these works carried out by other researchers, the proposed MIDNet18 model achieved a good accuracy of 98.5% in classification of brain tumors. Therefore, MIDNet18 performed significantly better than the AlexNet model and the other standard models that were discussed.
Our proposed MIDNet18 model outperformed the AlexNet model in brain tumor medical image classification of tumor and nontumor images. Our MIDNet18 model learned well and provided a high accuracy of more than 98% in binary classification, which is statistically significant (independent sample t-test P<0.05). MIDNet18 proves to be a high performer in all performance metrics with respect to the given dataset that was considered for this study.
The authors have no conflicts of interest in this manuscript.
Ramya Mohan was involved in novel MIDNet18 CNN architecture development, implementation, data collection, analysis and interpretation of data, and manuscript writing. Kirupa Ganapathy was involved in conceptualization, manuscript writing, document editing and critical review of manuscripts, and interpretation of the results. Rama A was involved in data collection, data analysis, and data validation.
The authors express their gratitude toward Saveetha School of Engineering, Saveetha Institute of Technical and Medical Sciences (formerly known as Saveetha University) for providing the necessary infrastructure to carry out this work successfully.
The authors are thankful for the following organizations for providing financial support that enabled us to complete this study.
Saveetha University
Saveetha Institute of Medical and Technical Sciences
Saveetha School of Engineering
1. Abdel Razek, AAK, Ahmed, A, Shehata, M, AbdelKhalek, A, Baky, KA, El-Baz, A., and Helmy, E. (2021). Clinical Applications of Artificial Intelligence and Radiomics in Neuro-Oncology Imaging. Insights into Imaging 12(1): 152. 10.1186/s13244-021-01102-6
2. Arshad, M, Khan, MA, Tariq, U, Armghan, A, Alenezi, F, Javed MY, Aslam, SM, and Kadry, S. (2021). A Computer-Aided Diagnosis System Using Deep Learning for Multiclass Skin Lesion Classification. Computational Intelligence and Neuroscience 2021: 9619079. 10.1155/2021/9619079
3. Bondy, ML, Scheurer, ME, Malmer, B, Barnholtz-Sloan, JS, Davis, FG, Il’yasova, D, Kruchko, C, et al. (2008). Brain Tumor Epidemiology: Consensus from the Brain Tumor Epidemiology Consortium. Cancer 113 (7 Suppl): 1953–1968. 10.1002/cncr.23741
4. Caruso, G, Merlo, L, and Caffo, M. (2014). Brief Introduction on Brain Tumor Epidemiology and State of the Art in Therapeutics. Innovative Brain Tumor Therapy. 2014: 1–15. 10.1533/9781908818744.1
5. Consortium, The Childhood Brain Tumor, The Childhood Brain Tumor Consortium, and Gilles, FH. (1991). The Epidemiology of Headache among Children with Brain Tumor. Journal of Neuro-Oncology. 10: 31–46. 10.1007/BF00151245
6. Gaballah, AM, Sarhan, AM, and Elshennawy, NM. (2021). Classification of Brain MRI Tumor Images Based on Deep Learning PGGAN Augmentation. Diagnostics. 11 (12). 10.3390/diagnostics11122343
7. Hang, ST, and Aono, M. (2017). Bi-Linearly Weighted Fractional Max Pooling. Multimedia Tools and Applications. 76: 22095–22117. 10.1007/s11042-017-4840-5
8. Hossain, MJ, Xiao, W, Tayeb, M, and Khan, S. (2021). Epidemiology and Prognostic Factors of Pediatric Brain Tumor Survival in the US: Evidence from Four Decades of Population Data. Cancer Epidemiology. 72: 101942. 10.1016/j.canep.2021.101942
9. Huang, J, Shlobin, NA, Lam, SK, and DeCuypere, M. (2021). Artificial Intelligence Applications in Pediatric Brain Tumor Imaging: A Systematic Review. World Neurosurgery. 157 (October): 99–105. 10.1016/j.wneu.2021.10.068
10. Ismael, MR. (2018). Hybrid Model: Statistical Features and Deep Neural Network for Brain Tumor Classification in MRI Images.
11. Johnson, KJ, Broholm, H, Scheurer, ME, Lau, CC, Hainfellner, JA, Wiemels, J, and Schwartzbaum, J. (2018). Advancing Brain Tumor Epidemiology–Multi-Level Integration and International Collaboration: The 2018 Brain tumor epidemiology consortium meeting report. Clinical Neuropathology. 37: 254–261. 10.5414/NP301148
12. Kaggle: Your Machine Learning and Data Science Community. n.d. Accessed December 17, 2021. https://www.kaggle.com/
13. Kang, J, Ullah, Z, and Gwak, J. (2021). MRI-Based Brain Tumor Classification Using Ensemble of Deep Features and Machine Learning Classifiers. Sensors 21 (6). 10.3390/s21062222
14. Kesav, N, and Jibukumar, MG. (2021). Efficient and Low Complex Architecture for Detection and Classification of Brain Tumor Using RCNN with Two Channel CNN. Journal of King Saud University - Computer and Information Sciences. 10.1016/j.jksuci.2021.05.008
15. Khan, AR, Khan, S, Harouni, M, Abbasi, R, Iqbal, S, and Mehmood, Z. (2021). Brain Tumor Segmentation Using K-Means Clustering and Deep Learning with Synthetic Data Augmentation for Classification. Microscopy Research and Technique. 84 (7): 1389–399. 10.1002/jemt.23694
16. Kremelberg, D. (2010). Practical Statistics: A Quick and Easy Guide to IBM® SPSS® Statistics, STATA, and Other Statistical Software. SAGE Publications, Incorporated. 10.4135/9781483385655
17. Liu, GS, Yang, A, Kim, D, Hojel, A, Voevodsky, D, Wang, J, Tong, CCL, et al. (2022). Deep Learning Classification of Inverted Papilloma Malignant Transformation Using 3D Convolutional Neural Networks and Magnetic Resonance Imaging. International Forum of Allergy & Rhinology. 10.1002/alr.22958
18. Lotlikar, VS, Satpute, N, and Gupta, A. (2021). Brain Tumor Detection Using Machine Learning and Deep Learning: A Review. Current Medical Imaging Reviews. 10.2174/1573405617666210923144739
19. Miller, KD, Ostrom, QT, Kruchko, C, Patil, N, Tihan, T, Cioffi, G, Fuchs, HE, et al. (2021). Brain and Other Central Nervous System Tumor Statistics, 2021. CA: A Cancer Journal for Clinicians. 71 (5): 381–406. 10.3322/caac.21693
20. Murray, N, and Perronnin, F. (2014). Generalized Max Pooling. 2014 IEEE Conference on Computer Vision and Pattern Recognition. 10.1109/CVPR.2014.317
21. Özcan, H, Emiroğlu, BG, Sabuncuoğlu, H, Özdoğan, S, Soyer, A, and Saygı, T. (2021). A Comparative Study for Glioma Classification Using Deep Convolutional Neural Networks. Mathematical Biosciences and Engineering: MBE. 18 (2): 1550–1572. 10.3934/mbe.2021080
22. Ranjbarzadeh, R, Kasgari, AB, Ghoushchi, SJ, Anari, S, Naseri, M, and Bendechache, M. (2021). Brain Tumor Segmentation Based on Deep Learning and an Attention Mechanism Using MRI Multi-Modalities Brain Images. Scientific Reports. 11 (1): 10930. 10.1038/s41598-021-90428-8
23. Saba, T, Abunadi, I, Sadad, T, Khan, AR, and Bahaj, SA. (2021). Optimizing the Transfer-Learning with Pretrained Deep Convolutional Neural Networks for First Stage Breast Tumor Diagnosis Using Breast Ultrasound Visual Images. Microscopy Research and Technique. 10.1002/jemt.24008
24. Tandel, GS, Tiwari, A, and Kakde, OG. (2021). Performance Optimisation of Deep Learning Models Using Majority Voting Algorithm for Brain Tumour Classification. Computers in Biology and Medicine. 135 (August): 104564. 10.1016/j.compbiomed.2021.104564
25. Tazin, T, Sarker, S, Gupta, P, Ayaz, FI, Islam, S, Khan, MM, Bourouis, S, Idris, SA, and Alshazly, H. (2021). A Robust and Novel Approach for Brain Tumor Classification Using Convolutional Neural Network. Computational Intelligence and Neuroscience. 2021 (December): 2392395. 10.1155/2021/2392395
26. Wang, H, Liu, C, Zhao, Z, Zhang, C, Wang, X, Li, H, Wu, H, et al. (2021). Application of Deep Convolutional Neural Networks for Discriminating Benign, Borderline, and Malignant Serous Ovarian Tumors From Ultrasound Images. Frontiers in Oncology. 11 (December): 770683. 10.3389/fonc.2021.770683
27. Wang, Y-W, Chii-Jen, C, Wang, T-C, Huang, H-C, Chen, H-M, Shih, J-Y, Chen, J-S, Huang, Y-S, Chang, Y-C, and Chang, R-F. (2021). Multi-Energy Level Fusion for Nodal Metastasis Classification of Primary Lung Tumor on Dual Energy CT Using Deep Learning. Computers in Biology and Medicine. 141 (December): 105185. 10.1016/j.compbiomed.2021.105185
28. Williams, S, Horsfall, HL, Funnell, JP, Hanrahan, JG, Khan, DZ, Muirhead, W, Stoyanov, D, and Marcus, HJ. (2021). Artificial Intelligence in Brain Tumour Surgery—An Emerging Paradigm. Cancers. 13 (19): 5010. 10.3390/cancers13195010