You are viewing a javascript disabled version of the site. Please enable Javascript for this site to function properly.
Go to headerGo to navigationGo to searchGo to contentsGo to footer
In content section. Select this link to jump to navigation

COVID-19 detection with X-ray images by using transfer learning

Abstract

Coronavirus is an infectious disease induced by extreme acute respiratory syndrome coronavirus 2. Novel coronaviruses can lead to mild to serious symptoms, like tiredness, nausea, fever, dry cough and breathlessness. Coronavirus symptoms are close to influenza, pneumonia and common cold. So Coronavirus can only be confirmed with a diagnostic test. 218 countries and territories worldwide have reported a total of 59.6 million active cases of the COVID-19 and 1.4 million deaths as of November 24, 2020. Rapid, accurate and early medical diagnosis of the disease is vital at this stage. Researchers analyzed the CT and X-ray findings from a large number of patients with coronavirus pneumonia to draw their conclusions. In this paper, we applied Support Vector Machine (SVM) classifier. After that we moved on to deep transfer learning models such as VGG16 and Xception which are implemented using Keras and Tensor flow to detect positive coronavirus patient using X-ray images. VGG16 and Xception show better performances as compared to SVM. In our work, Xception gained an accuracy of 97.46% with 98% f-score.

1Introduction

Coronavirus is a contagious virus that affects respiratory infection and is widely contagious from person to person. The World Health Organization (WHO) noted that many cases of respiratory infection were clinically similar to viral pneumonia, breathlessness, mild fever and cough on December 31, 2019 [1]. The Novel coronavirus (2019-nCoV) was identified in Wuhan providence, China. It refers to a class of viruses including Middle East Respiratory Syndrome (MERS) and Extreme Acute Respiratory Syndrome (SARS) [2]. Coronaviruses may often cause diseases such as pneumonia or bronchitis in the lower respiratory tract. This is much more severe in people with poor immune systems, heart, kidney disease and liver disease [3]. COVID-19 confirmed cases, death cases and recovery cases in Brazil are 6.08, 0.17 and 5.44 million respectively as of November 24, 2020. In India number of coronavirus confirmed cases (9.17 million) is higher than Brazil but the number of death cases (0.13 million) is very less [4, 30–33].

The detection of COVID-19 should be confirmed by gene sequences for respiratory or blood samples as a primary indicator of reverse transcription polymerase chain reaction (RT-PCR), as per the Chinese government. RT-PCR sensitivity is very low. So that it is very difficult to diagnose many COVID- 19 patients [25, 34–35]. The diagnosis currently involves anyone who discovers the popular pneumonia of a COVID-19 chest scan rather than people waiting for positive tests on the virus. This method will allow doctors to quickly isolate and treat the positive COVID-19 patients.

Actually, the COVID-19 check is a difficult task because the diagnostic system is unavailable in certain places. We have to rely on other diagnostic methods to overcome the restricted supply testing kits for COVID-19. Since COVID-19 targets our respiratory tract’s epithelial cells, we may use X-rays to monitor the condition of a patient’s lungs. Doctors used X-ray devices to scan infected body parts for tumors, lung infections, pneumonia and fractures. Computer tomography (CT) is an imaging technique that uses different devices to scan and produce accurate images of areas within the body. It is more advanced as compared to X-ray. Although most hospitals have X-ray scanning technology, it may be essential to use X-rays of the chest to examine COVID-19. As compared to CT scan technology X-ray scan is better. X-ray scan is cheaper, easier, faster and less dangerous to the human body. The disadvantage of X-rays analysis requires a radiology specialist to detect COVID-19 positive patient within short span of time. Some patients may also have symptoms of coronavirus but still they are COVID-19 negative. So, an automated model must be built to save valuable time for medical practitioners to detect COVID-19 patients.

Deep transfer learning has extensive applications in medical image analysis. Various researches have published many research articles to detect brain tumor, breast cancer, and respiratory illness. With the help of deep transfer learning the pediatric pneumonia is identified through chest X-ray images, nature of pulmonary nodules identified through CT images and cystoscopic image processing extracted through videos etc [5]. In this manuscript, we use SVM, VGG16 and Xception pre-trained models on chest X-ray images to detect COVID-19 positive patients.

The rest of the article is arranged as follows. Literature survey is presented in section 2, SVM and deep transfer learning models are described in section 3, section 4 contains the performance matrices and conclusion followed by future works is presented in section 5.

2Related works

In this portion, we presented a literature survey on various methodologies proposed by researchers to detect COVID-19 patients.

Shuai Wang et al. [6] suggested a deep learning technique that would retrieve the graphic features of COVID-19, thereby saving crucial time for disease control. They use 1119 COVID-19 positive CT images along with pneumonia images. They modified the existing inception deep learning approaches to detect COVID-19 patients. They got the accuracy as 79.3% and 89.5% for external and internal validations respectively.

Fei Shan et al. [7] proposed deep learning model named “VB-Net” to segment infected COVID-19 CT scans regions. Their model is trained and validated using 249, 300 COVID19 positive patients respectively. Authors have implemented a human-in-the-loop (HITL) procedure to enable radiologists to improve automated annotation. Their model produced dice similarity coefficients between automatic and manual segmentations of 91.6% ±10.0%.

Xiaowei Xu et al. [8] used 3-D deep CNN model from pulmonary CT image set to segment infection regions as Influenza-A, positive COVID-19 and not relevant to infection groups. They use location attention classification model and Bayesian function to calculate the confidence scores. Their model produced the accuracy as 86.7%.

Ophir Gozes et al. [9] developed automatic CT image analysis systems focused on artificial intelligence (AI) to diagnose, analyze and track coronavirus and to differentiate negative COVID-19 patients. Classification results for active COVID-19 versus negative COVID-19 cases were achieved by 0.996 Area under Curve thoracic CT studies with 98% specificity and 94% sensitivity.

Ali Narin et al. [10] proposed InceptionResNetV2, InceptionV3 and ResNet50 model for the identification of positive coronavirus patients utilizing chest X-ray datasets. ResNet50 model offers the best accuracy value as compared to other models.

Prabira et al. [11] proposed coronavirus identification model on X-ray images using different transfer learning models with SVM. They retrieved the deep CNN model features and fed each individually to the SVM classifier. Their proposed COVID-19 classification model was 95.38% accurate.

For pulmonary tuberculosis (PTB), Wei Wu et al. [12] recommended a deep CNN system to automatically generate a CT reports for diagnostic study. They implemented 3D deep learning models to examine lungs and could use the advantages of 3D CT imaging technology to accurately detect different regions and forms of lesions. Among them, v-net backbone with inception-resnet blocks provided the best results for both detection accuracy and classification.

CT imaging features of 2019 novel coronavirus [13–16] are beneficial for the identification and treatment of COVID-19 by the radiologist.

3Materials and methods

In this portion we discuss the traditional classification model (SVM) and deep transfer learning models (VGG16 & Xception).

3.1Traditional classification (SVM)

For machine learning classification tasks basically Support Vector Machine (SVM) used. SVM is focused on supervised learning mechanism that classifies points to one of the two disjoint planes. Vapnik [17] proposed the SVM for binary classification. The objective of SVM is to obtain the optimum hyper plane via f (w, x)=w · x + b to distinguish two classes in a particular dataset by maximizing the margin, with features x ∈ Rm.

SVM learns certain parameters W by resolving the problem of optimization mentioned in Equation 1.

(1)
min1pWTW+Ci=1pmax(0,1-yi(WTXi+b))

Where WT W = manhattan norm, y=actual label, C = penalty parameter, WT X + b=predictor function. Equation 1 is also called as standard hinge loss (L1-SVM). L2-SVM (Squared hinge loss) [18] provides more stable results represented in equation 2. ∥W ∥ 2 represented as Euclidean norm.

(2)
min1p||W||22+Ci=1pmax(0,1-yi(WTXi+b))2

SVM does not do so well whenever the data set has much more noise, i.e. the target classes overlap. So we use deep transfer learning models.

3.2VGG16

VGG 16 was proposed by Karen Simonyan et al. [19] in 2014. VGG-16 comes from the fact that it has 16 layers. VGG16 contains different layers such as 3 dense layers, 2 fully connected layers, 5 max pooling layers and 13 convolutional layers. All total it contains 16 weight layers out of 21 layers.

1st and 2nd layers:

AlexNet’s input is a 224x224x3 RGB image that passes through first and second convolutionary layers with 64 filters with size 3x3 and the same pooling with a 14 phase. The dimensions of the images shift to 224x224x64. The VGG16 then applies a maximum layer of pooling or a sub-sampling layer with a filter size 3x3 and a phase of 2. Dimensions of the resulting image are reduced to 112x112x64.

3rd and 4th layers:

There are two convolutionary layers with 128 size 3x3 feature maps and a phase of 1. Then again there is a maximum layer of pooling with filter size 3x3 and a phase of 2. This layer is the same as the previous pooling layer except it has 128 feature maps which reduce the performance to 56x56x128.

5th and 6th layers:

5th and 6th layers are convolutionary filter size layers 3x3 and one phase. Both employed 256 maps of functions. A maximum pooling layer with filter size 3x3, a stage of 2 and 256 feature maps follow the two convolution layers.

7th-12th Layer:

These layers contain two sets of three convolutionary layers and a maximum pooling layer. All convolution layers have 512 size 3x3 filters and one phase. The final size is trimmed to 7x7x512.

13th Layer:

A fully connected layer (FCL) with a 25088 feature maps each size 1x1 flattens the overall output layer.

14th and 15th layers:

These layers contain FCL with 4096 units.

Output Layer:

Output layer is called as softmax output layer ŷ with 1000 possible values.

The detail VGG16 architecture is illustrated in Fig. 1.

Fig. 1

Basic layout of VGG16 architecture showing its 5 Max Pooling layers, 13 convolutional layers and 2 fully connected layers.

Basic layout of VGG16 architecture showing its 5 Max Pooling layers, 13 convolutional layers and 2 fully connected layers.

Challenges of VGG 16:

VGG16 model is very slow to train the dataset. The imageNet weights trained by VGG-16 are 528 MB. Therefore, a lot of disk space and bandwidth are required to make it ineffective.

3.3Xception

Several important concepts used to develop Xception [20–22] architecture, including depth-wise separable convolution layer, convolutional layer, residual connections and inception module. Xception is the acronym for “extreme inception”. The design of Xception has 36 convolutionary layers which form the network’s extractor base. Xception is an extension of Inception Architecture by deeply separable convolutions for standard Inception modules. Speed and accuracy of Xception model is better as compared to Inception, ImageNet, Inception V3, ResNet-101, ResNet-50, VGGNet. Model. Xception model contains 3 major sections: entry, middle and exit flow. The data goes first through the input flow, then through the eight-fold middle flow, and finally through the exit flow. Batch normalization is accompanied by all the Convolution layers. Xception replaced the conventional starting block by expanding it and changing the various (1x1, 5x5, 3x3) spatial dimensions with a (3x3) single dimension preceded by a (1x1) transition to maintain computational burden.

Xception building block is depicted Fig. 2. Equations 3 and 4 represents the mathematical modeling of Xception.

Fig. 2

Xception building block architecture.

Xception building block architecture.

(3)
fl+1k(p,q)=x,yflk(x,y)·elk(u,v)
(4)
Fl+2k=gc(Fl+2k,kl+1)

By decoupling spatial and feature-map correlation, Xception makes a network computationally effective. Using 1x1 convolutions it maps the transformed output to low dimensional embedding. Then it transforms spatially n times, in which n is the cardinality, defining width that defines the number of transformations.

In Equation (4), kl is a kth kernel of lth layer with one depth that is spatially transformed across the kth feature-map flk . Kernel and feature-map spatial indices are represented as (x, y) and (u, v) respectively.

The number of feature-maps is same as of kernels K in depth wise separable convolution. Although in the case of standard convolutionary layers, the number of kernels is not depending on previous layer. On the other hand kl +1 is kth kernel of (1x1) spatial dimension for l + 1th layer. Output feature-maps [ Fl+1,1Fl+1,kFl+1,k ] of lth layer, l + 1th input layer performed as depth wise convolution.

Xception enables computation by translating each function map separately around spatial axes, followed by point-specific convolution (1x1 convolutions) for cross-channel correlation. In deep CNN; the number of filters used in Xception is same as the transformation segments. But three transformation segments are used by inception block. Although Xception’s transformation policy does not decrease the parameter numbers, it ensures learning more effective and performs better results. Xception building block is represented in Fig. 2.

Advantages of Xception model

Xception model is lighter due to the less number of connections. It uses depth wise separable convolutions. This model is stronger and robust than the Inception model.

4Proposed work

We applied three different algorithms on our covid-19 dataset, which consists of X-RAY images of chest for the prediction of COVID-19, well known VGG16 and Xception transfer learning models have been trained and tested on the available images. Working of VGG16 & Xception models for the detection of COVID-19 positive and negative patients are depicted in Fig. 3.

Fig. 3

Working principle of VGG16 & Xception pre-trained model for detection of positive and negative coronavirus patients.

Working principle of VGG16 & Xception pre-trained model for detection of positive and negative coronavirus patients.
Proposed Algorithm: Detection of Positive and Negative Patients
Step 1: Take X-ray images (COVID-19 &Normal chest images) as input
Step 2: Load pre-trained models like VGG16 and Xception
Step 3: In training phase of VGG16, we use GlobalAveragePooling2D followed by FCL with dropout 0.5, FCL with ReLU 98 and finally FCL with softmax as output layer
Step 4: In training phase of Xception, we use FCL with dropout 0.5 followed by FCL with ReLU 98 and finally FCL with softmax as output layer
Step 5: Detect the COVID-19 positive and negative patients as output
  Global Average Pooling 2D applies average pooling on the spatial dimensions until each spatial dimension is one. FCL with dropout 0.5 is used to prevent overfitting. FCL with ReLU 98 is used to improve neural networks by speeding up training. FCL with softmax function is used to normalize the outputs between 0 and 1 range.

5Real time discussion

5.1Experimental setup

The entire experiment was conducted on Google Colaboratory, using Python3 engine and GPU as backend with system RAM 12 GB. Tensor flow 1x is used. For image processing, we loaded the image as color having 3 RGB channels and converted the images to numpy array and rescaled by dividing by 255.The measurement of performance of each pre-trained model (SVM, VGG16, and Xception) is measured in terms of Recall, Precision, F1-Score and Specificity represented in Table 2. The batch size is taken as 32 where the learning rate used in Xception 1e – 2, Epochs is set to be 100. All the models are trained using Adam optimizer.

5.2Dataset description

We collect COVID-19 X-ray image dataset from GitHub uploaded by Joseph Paul Cohen [23], Postdoctoral Fellow, Mila, Universit e´ de Montr e´ al. We obtained normal chest X-ray images from Kaggle repository uploaded by Paul Mooney [24], Developer Advocate at Kaggle, Boulder, CO, USA. Active COVID-19 and normal patients X-ray images are illustrated in Fig. 4. These images are being used to train the Tensor Flow and Keras deep learning model to automatically determine if an individual has COVID-19. Number of images utilized for training, testing and validating are presented in Table 1.

Fig. 4

(a) positive (i.e., infected) COVID-19 X-ray images (b) negative samples.

(a) positive (i.e., infected) COVID-19 X-ray images (b) negative samples.
Table 1

Dataset description

COVID-19Coronavirus Positive ImagesCoronavirus Negative Images
Train6767
Valid2222
Test1515
Total104104

5.3Results and outcomes

The accuracy and loss graphs of these two models are given in Figs. 5 and 6 respectively.

Fig. 5

Accuracy graph for the training phase of VGG16 v/s Xception.

Accuracy graph for the training phase of VGG16 v/s Xception.
Fig. 6

Loss Graph for the Training phase of VGG16 v/s Xception.

Loss Graph for the Training phase of VGG16 v/s Xception.

It can be interpreted from the plot that Xception learns faster than VGG16 from the Fig. 5. Xception has obtained greater Accuracy that VGG16 for the same number of Epochs i.e. 100. To save time and resources early stopping and model checkpoints have been used to save the best model and stop training if there is no significant improvement in the accuracy. The ImageNet weights has been initialized for both the models. However Xception stands out in this race. From Fig. 6, the loss for both the models has been depicted, The VGG16 have initially very high loss rate but there is a gradual decline in the loss whereas Xception have a smooth descent. Almost for both the models, the loss almost approaches within the range <  0.1.

Performance metrics

The measurement of performance of each pre-trained model (SVM, VGG16, Xception) is measured in terms of Recall, Precision, F1-Score, Sensitivity, Specificity and the results are represented in Table 2. Confusion Matrix of all the models is illustrated in Figs. 79. Overall performances of models are represented in Fig. 10.

Table 2

Represents performance metrics of SVM, VGG16 and Xception Models

ModelPrecisionRecallF1-ScoreSpecificity
SVM01.000.910.950.913043
10.921.000.96
VGG1600.960.960.960.956521
10.950.950.95
Xception01.000.960.980.956521
10.961.000.98
Fig. 7

Confusion matrix of SVM.

Confusion matrix of SVM.
Fig. 8

Confusion matrix of VGG16.

Confusion matrix of VGG16.
Fig. 9

Confusion matrix of Xception.

Confusion matrix of Xception.
Fig. 10

Performance of VGG16, Xception and SVM Models.

Performance of VGG16, Xception and SVM Models.

Performance Metrics used are:

  • 1. Precision=TruePositive(TP)TruePositive+FaisePositive(FP)

  • 2. Recall=TruePositiveTruePositive+FaiseNegative(FN)

  • 3. F1-Score=2x(PrecisionxRecall)(Precision+Recall)

  • 4. Specificity=TrueNegative(TN)TrueNegative+FaisePositive

Confusion Matrix of SVM, VGG16, Xception

From confusion matrix, displayed as Fig. 7, we can see out of 45 samples

21 samples have been correctly classified as positive

22 images have been correctly classified as negative

2 images have been classifies as positive but actually is negative

0 images have been classifies as negative but actually is positive

From confusion matrix, displayed as Fig. 8, we can see out of 45 samples

22 samples have been correctly classified as positive

21 images have been correctly classified as negative

1 images have been classifies as positive but actually is negative

1 images have been classifies as negative but actually is positive

From confusion matrix, displayed as Fig. 9, we can see out of 45 samples

22 samples have been correctly classified as positive

22 images have been correctly classified as negative

1 images have been classifies as positive but actually is negative

0 images have been classifies as negative but actually is positive

Overall performance

It can be observed that among the three models, Xception model is having 97.46% of accuracy as compared to SVM and VGG16. Accuracy can be evaluated as:

Accuracy=TP+TNTP+TN+FP+FN

Table 3

Comparative analysis of proposed work with existing work

AuthorsChest CT/X-raysModel or MethodologyAccuracy (%)
Shuai Wang et al. [6]CTEnhanced Inception deep transfer-learning model89.5%
Xiaowei Xu et al. [8]CT3-dimensional deep learning model86.7%
Xu et al. [26]CTLocation Attention with ResNet86.7%
Zheng et al. [27]CT3D Deep Network with UNet90.8%
Narin et al. [10]X-rayResNet-50 and Deep CNN97%
Prabira et al. [11]X-rayDeep CNN models with SVM95.38%
Sethy et al. [28]X-raySVM with ResNet5095.38%
Abbas et al. [29]X-rayDeTraC (Decompose, Transfer, and Compose) Deep CNN95.12%
Proposed ModelX-rayVGG16 &Xception96.54% and 97.46%

6Conclusion

Early diagnosis and treatment of the COVID-19 patients are necessary to prevent the disease from spreading to others. In this paper, we have applied 3 models namely SVM, VGG16 and Xception. SVM being a traditional machine learning algorithm that works on the maximum margin construction concept for the case of 1 dimensional classification. This comes with some limitations. To resolve this issue, we made use of VGG16 and Xception, both of which are deep transfer learning models, to improve the efficiency of prediction of COVID-19 positive patients. The Xception transfer learning model yields the maximum accuracy of 97.46% with 98% f-score. In future, we plan to detect COVID-19 by using neutrosophical principles with different deep convolutional neural network models by utilizing maximum images in the dataset.

References

[1] 

(2020). Novel Coronavirus – China WHO Data. [Online] Available: https://www.who.int/csr/don/12-january-2020-novel-coronavirus-china/en/

[2] 

(2020). Coronaviruses and Acute Respiratory Syndromes (COVID-19, MERS, AND SARS) – MSD manual Data. [Online] Available: https://www.msdmanuals.com/professional/infectious-diseases/respiratory-viruses/coronaviruses-and-acute-respiratory-syndromes-covid-19,-mers,-and-sars

[3] 

(2020). Common Human Coronaviruses – Centers for Disease Control and Prevention Data. [Online] Available: https://www.cdc.gov/coronavirus/general-information.html

[4] 

(2020) COVID-19 Coronavirus Pandemic – worldometer Data. [Online] Available: https://www.worldometers.info/coronavirus/#countries

[5] 

Choe, Jooae, Sang Min Lee, Kyung-Hyun Do, Gaeun Lee, June-Goo Lee, Sang Min Lee and Joon Beom Seo, Deep Learning–based Image Conversion of CT Reconstruction Kernels Improves Radiomics Reproducibility for Pulmonary Nodules or Masses, Radiology 292: (2) ((2019) ), 365–373. DOI: 10.1148/radiol.2019181960

[6] 

Wang, Shuai, Bo Kang, Jinlu Ma, Xianjun Zeng, Mingming Xiao, Jia Guo, Mengjiao Cai et al. A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19), MedRxiv ((2020) ). DOI: 10.1101/2020.02.14.20023028

[7] 

Shan, Fei, Yaozong Gao, Jun Wang, Weiya Shi, Nannan Shi, Miaofei Han, Zhong Xue and Yuxin Shi, Lung infection quantification of covid-19 in ct images with deep learning, arXiv preprint arXiv:2003.04655 ((2020) ).

[8] 

Xu, Xiaowei, Xiangao Jiang, Chunlian Ma, Peng Du, Xukun Li, Shuangzhi Lv, Liang Yu et al. A deep learning system to screen novel coronavirus disease 2019 pneumonia, Engineering ((2020) ). DOI: 10.1016/j.eng.2020.04.010

[9] 

Gozes, Ophir, Maayan Frid-Adar, Hayit Greenspan, Patrick D. Browning, Huangqi Zhang, Wenbin Ji, Adam Bernheim and Eliot Siegel, Rapid AI Development Cycle for the Coronavirus (COVID-19) Pandemic: Initial Results for Automated Detection & Patient Monitoring using Deep Learning CT Image Analysis, arXiv preprint arXiv:2003.05037 ((2020) ).

[10] 

Narin, Ali, Ceren Kaya and Ziynet Pamuk, Automatic Detection of Coronavirus Disease (COVID-19) Using X-ray Images and Deep Convolutional Neural Networks, arXiv preprint arXiv:2003.10849 ((2020) ).

[11] 

Sethy, Prabira Kumar and Santi Kumari Behera, Detection of coronavirus disease (Covid-19) based on deep features, Preprints 2020030300: ((2020) ), 2020. DOI: 10.20944/preprints202003.0300.v1

[12] 

Wu, Wei, Xukun Li, Peng Du, Guanjing Lang, Min Xu, Kaijin Xu and Lanjuan Li, A Deep Learning System That Generates Quantitative CT Reports for Diagnosing Pulmonary Tuberculosis, arXiv preprint arXiv:1910.02285 ((2019) ).

[13] 

Chung, Michael, Adam Bernheim, Xueyan Mei, Ning Zhang, Mingqian Huang, Xianjun Zeng, Jiufa Cui et al. CT imaging features of 2019 novel coronavirus (2019-nCoV), Radiology 295: (1) ((2020) ), 202–207. DOI: 10.1148/radiol.2020200230

[14] 

Yoon, Soon Ho, Kyung Hee Lee, Jin Yong Kim, Young Kyung Lee, Hongseok Ko, Ki Hwan Kim, Chang Min Park and Yun-Hyeon Kim, Chest Radiographic and CT Findings of the 2019 Novel Coronavirus Disease (COVID-19): Analysis of Nine Patients Treated in Korea, Korean Journal of Radiology 21: (4) ((2020) ), 494–500. DOI: 10.3348/kjr.2020.0132

[15] 

Yuan, Mingli, Wen Yin, Zhaowu Tao, Weijun Tan and Yi Hu, Association of radiologic findings with mortality of patients infected with 2019 novel coronavirus in Wuhan, China, PloS One 15: (3) ((2020) ), e0230548. DOI: 10.1371/journal.pone.0230548

[16] 

Hosseiny, Melina, Soheil Kooraki, Ali Gholamrezanezhad, Sravanthi Reddy and Lee Myers, Radiology perspective of coronavirus disease 2019 (COVID-19): lessons from severe acute respiratory syndrome and Middle East respiratory syndrome, American Journal of Roentgenology ((2020) ), 1–5. DOI: 10.2214/AJR.20.22969

[17] 

Cortes, Corinna and Vladimir Vapnik, Support-vector networks, Machine Learning 20: (3) ((1995) ), 273–297.

[18] 

Tang, Yichuan, Deep learning using linear support vector machines, arXiv preprint arXiv:1306.0239 ((2013) ).

[19] 

Simonyan, Karen and Andrew Zisserman, Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv: 1409.1556, ((2014) ).

[20] 

Jinsakul, Natinai, Cheng-Fa Tsai, Chia-En Tsai and Pensee Wu, Enhancement of deep learning in image classification performance using xception with the swish activation function for colorectal polyp preliminary screening, Mathematics 7: (12) ((2019) ), 1170. DOI: 10.3390/math7121170

[21] 

Chollet, François, Xception: Deep learning with depth wise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1251–1258. (2017) . DOI: 10.1109/CVPR.2017.195

[22] 

Khan, Asifullah, Anabia Sohail, Umme Zahoora and Aqsa Saeed Qureshi, A survey of the recent architectures of deep convolutional neural networks. arXiv preprint a Xiv:1901.06032 ((2019) ). DOI: 10.1007/s10462-020-09825-6

[23] 

(2020) 2019 Novel Coronavirus COVID-19 (2019-nCoV) Data Repository by Johns Hopkins CSSE. [Online] Available: https://github.com/CSSEGISandData/COVID-19

[24] 

(2018) Chest X-Ray Images (Pneumonia) by Paul Mooney Kaggle. [Online] Available: https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia

[25] 

Zhang, Wei, Imaging changes of severe COVID-19 pneumonia in advanced stage, Intensive Care Medicine ((2020) ), 1–3. DOI: 10.1007/s00134-020-05990-y

[26] 

Xu, X., X. Jiang, C. Ma, P. Du, X. Li, S. Lv, L. Yu, Y. Chen, J. Su and G. Lang, Deep learning system to screen coronavirus disease 2019 pneumonia. arXiv 2020”. arXiv preprint arXiv:2002.09334. DOI: 10.1007/s10489-020-01714-3

[27] 

Zheng, Chuansheng, Xianbo Deng, Qing Fu, Qiang Zhou, Jiapei Feng, Hui Ma, Wenyu Liu and Xinggang Wang, Deep learning-based detection for COVID-19 from chest CT using weak label. medRxiv ((2020) ). DOI: 10.1101/2020.03.12.20027185

[28] 

Sethy, Prabira Kumar and Santi Kumari Behera, Detection of coronavirus disease (Covid-19) based on deep features, Preprints 2020030300: ((2020) ), 2020. DOI: 10.20944/preprints202003.0300.v1

[29] 

Abbas, Asmaa, Mohammed M. Abdelsamea and Mohamed Medhat Gaber, Classification of COVID-19 in chest X-ray images using DeTraC deep convolutional neural network, arXiv preprint arXiv:2003.13815 ((2020) ). DOI: 10.1101/2020.03.30.20047456

[30] 

Sánchez, Daniela et al., Comparison of Neural Network Models Applied to Human Recognition, Soft Computing Applications, Proceedings of the 8th International Workshop Soft Computing Applications (SOFA 2018), 1: ((2018) ), 130–142.

[31] 

Negirla, Paul-Onut et al., Mobile Robot Platform for Studying Sensor Fusion Localization Algorithms, Soft Computing Applications, Proceedings of the 8th International Workshop Soft Computing Applications (SOFA 2018), 1: ((2018) ), 317–326.

[32] 

Ali Kadhum Idrees and Wathiq Laftah Al-Yaseen, Distributed genetic algorithm for lifetime coverage optimisation in wireless sensor networks, International Journal of Advanced Intelligence Paradigms 18: (1) ((2021) ), 3–24.

[33] 

Rasim M. Alguliyev, Fargana J. Abdullayeva and Sabira S. Ojagverdiyeva, Protecting children on the internet using deep generative adversarial networks, International Journal of Computational Systems Engineering 6: (2) ((2020) ), 84–90.

[34] 

Malik H. , Almutairi A. and Alotaibi M.A. , Power quality disturbance analysis using data-driven EMD-SVM hybrid approach, Journal of Intelligent & Fuzzy Systems 1–10 ((2021) ).

[35] 

Bai H. , Zhang H. and Wang Q. , Dual discriminative auto-encoder network for zero shot image recognition, Journal of Intelligent & Fuzzy Systems 1–12 ((2021) ).