Chapter 6: Hybrid computer-aided classification system design using end-to-end Pre-trained CNN-based deep feature extraction and PCA-SVM classifier for chest radiographs

Abstract

This chapter covers the exhaustive description of the experiment carried out for the design of a hybrid CAC system using end-to-end Pre-trained GoogLeNet CNN model as a deep feature extractor and PCA-SVM classifier for chest radiographs. It explains in detail the concepts of feature dimensionality reduction and the steps involved in implementing PCA and SVM classifier. The code snippets of the different experiments aim at giving a better understanding to the programmatic implementation of designing these CAC systems.

Keywords

Hybrid CAC system; DAG network; Deep feature extraction; GoogLeNet; Feature selection; Correlation-based feature selection; Feature dimensionality reduction; Principal component analysis; Machine learning; SVM classifier

6.1: Introduction

This chapter covers the exhaustive description of the experiments carried out for the design of a hybrid computer-aided classification (CAC) system using an end-to-end Pre-trained GoogLeNet convolution neural network (CNN) model as deep feature extractor and principal component analysis support vector machine (PCA-SVM) classifier for chest radiographs. It explains in detail the concepts of feature dimensionality reduction and the steps involved in implementing PCA and SVM classifier. The code snippets of the experiment aim at giving a better understanding to the programmatic implementation of designing this CAC system.

6.2: Experimental workflow

The experimental workflow followed for analyzing the performance of a CAC system designed for binary classification of chest radiographs using deep feature extraction, GoogLeNet CNN model, and PCA-SVM is shown in Fig. 6.1.

Fig. 6.1
Fig. 6.1 Experimental workflow of CAC system designed using deep feature extraction, GoogLeNet CNN model, and PCA-SVM classifier.

6.3: Deep feature extraction

The process of extracting the features of an image from the deep layers of a CNN is referred to as deep feature extraction, and the features extracted are called deep features. This process involves the steps of providing the input data to the Pre-trained CNN, and then the respective activation values from the fully connected layer usually present at the end of the network or pooling layer are obtained. The process of deep feature extraction has been discussed in detail in Chapter 5. In the present work, the deep features are extracted from the best fine-tuned model decided on the basis of the classification results of Experiments 1–3, discussed in Chapter 4, consequentially forming a set of deep features that are extracted from the best performing CNN model. This feature set is referred to as a deep feature set (DFS).

The following code snippets show the process of feature extraction and feature map visualization of the extracted features. Code Snippet 6.1 shows the syntax to load the GoogLeNet CNN model trained in Experiment 3, discussed in Chapter 4, to extract the features from its pooling layer.

Code Snippet 6.1

Loading the saved GoogLeNet Pre-trained CNN model

%%loading the saved GoogLeNet Pre-trained CNN model%%load(‘Exp3_GoogLeNet’);

Code Snippet 6.2 shows the extraction of training and testing features from the pooling layer of the GoogLeNet CNN model after it has been loaded into the MATLAB workspace.

Code Snippet 6.2

Feature extraction from the GoogLeNet Pre-trained CNN model

%%feature extraction%% layer = 'pool5'; featuresTrain  =  activations(Exp3_GoogLeNet,imds_train,layer,'OutputAs','rows');featuresTest  =  activations(Exp3_GoogLeNet,imds_test,layer,'OutputAs','rows');

Code Snippet 6.3 shows the syntax of writing the features extracted from the pooling layer to Microsoft Excel sheets. These .xlsx files or .csv files are then used as input to the feature selection methods and machine learning-based classifiers.

Code Snippet 6.3

Saving the feature extracted from the GoogLeNet Pre-trained CNN model

%%saving the features in a excel sheet%%xlswrite('H:DataChestX-rayFeatureExtractionExp6_Train.xlsx',featuresTrain);xlswrite('H:DataChestX-rayFeatureExtractionExp6_Test.xlsx',featuresTest);

6.4: Feature selection and dimensionality reduction

The process of feature selection and different methods of feature selection have been discussed in the previous chapter along with the details of GoogLeNet as a deep feature extractor and the feature visualization.

6.4.1: Correlation-based feature selection

Correlation-based feature selection (CFS) is a filter-based feature selection technique that aims at extracting the best and most optimal set of features, which are nothing but a subset of the original set of features, on the basis of their correlation values. This technique is widely used as an attempt to select the optimal features from the original feature sets [112]. The CFS follows the basic idea that the features are not correlated to each other but have higher correlations to the class to which they belong; here, the reduced feature set (RFS) consists of the features that are uncorrelated to each other but have high correlation to their respective classes, either Normal or Pneumonia. A detailed explanation of CFS is given in Chapter 5.

The DFS contains the features extracted from the pooling layer of the GoogLeNet CNN, that is, the global average pool (GAP) layer, which when subjected to the CFS, results in an RFS. The DFS consists of 1024 features extracted from the GAP layer, and after CFS, the RFS consists of 103 features.

6.4.2: PCA-based feature dimensionality reduction

In the present work, PCA is used as a dimensionality reduction technique for the deep feature space, which is extracted from the Pre-trained GoogLeNet CNN model in the form of a DFS. This DFS, after further application of CFS, forms an RFS. The PCA helps in finding optimal principal components (PCs) that are highly useful in the classification task [1316]. The steps involved for the implementation of PCA are given in Fig. 6.2.

Fig. 6.2
Fig. 6.2 Steps in principal component analysis. PC, principal component.

The total resultant optimal PCs that need to be taken into consideration for the job of classification are obtained by repetition of the experiments in an iterative manner. In this way, the PCs are calculated empirically mainly by stepping through the few initial PCs that primarily lie in the range of 2–15 in an attempt to build a model.

6.5: SVM classifier

The SVM classifier is a member of the supervised machine learning algorithms class. The basic concept that the SVM works on is decision boundaries. The SVM applies functions called kernels that perform the mapping of the training data to a feature space of a higher dimensionality [1721]. The training data, which is mainly nonlinear in nature, is taken from the input space. Some of the most common kernels used in SVM are: (a) polynomial kernel, (b) radial basis function or Gaussian radial basis function, and (c) sigmoid kernel. The present work implements the SVM classifier using the LibSVM library [22], and the kernel used is the Gaussian radial basis function kernel. Among the multiple steps involved in implementing SVM, the most crucial step is attaining a good, generalized performance. The parameters represented by C and γ play a major role in achieving a good and well-generalized performance of the SVM classifier. Here the parameter C is responsible for the regularization of the result; hence, it is called the regularization parameter. Similarly, the parameter γ is associated with the kernel performance; hence it is called the kernel parameter. The main aim of regularization parameter C is to maximize the margin, which is the distance between the support vectors, while keeping the error of training as low as possible. In the present work, a k-fold cross validation is carried out, where k = 10 on the training dataset. Here each combination of (C, γ) is chosen, such that:

C{24,23215}

si1_e

γ=212,21124

si2_e

The grid search procedure followed in the parameter space aims at giving the optimum values of the regularization parameter C and the kernel parameter γ, such that the training accuracy is maximum at the optimum value of (C, γ) [2327]. The extracted feature values are normalized in the range [0,1] by using the min-max normalization algorithm. This normalization is performed in order to dodge any bias that could be caused by the presence of unbalanced feature values. The SVM has been widely used in classification of medical images [2839]. The steps followed in SVM are given in Fig. 6.3.

Fig. 6.3
Fig. 6.3 Support vector machine classifier.

6.6: Experiment and result

Experiment 6: Designing a hybrid CAC system for chest radiographs using deep feature extraction, GoogLeNet, and PCA-SVM classifier

From the results of Experiments 1–3, as described in Chapter 4, it can be seen that the GoogLeNet CNN model attains the highest accuracy (90.00%). Hence, for this experiment to evaluate the performance of designing a hybrid CAC system for chest radiographs using GoogLeNet, deep feature extraction, and PCA-SVM classifier, the GoogLeNet CNN is used as the deep feature extractor. The features from the GAP layer of GoogLeNet CNN are extracted forming a DFS of 1024 features. This DFS is reduced to 103 uncorrelated features resulting in the formation of an RFS by CFS. This RFS acts as an input to the PCA-SVM classifier for the classification of chest radiographic images. The result of the performance evaluation of the hybrid CAC system designed using deep feature extraction, GoogLeNet CNN model, and PCA-SVM classifier is shown in Table 6.1.

Table 6.1

Performance evaluation of CAC system designed for chest radiographs using deep feature extraction, GoogLeNet CNN, and PCA-SVM classifier.
Network/classifierConfusion matrixAccuracy (%)ICA_Normal (%)ICA_Pneumonia (%)
NormalPneumonia
GoogLeNet/PCA-SVMNormal48291.0096.0086.00
Pneumonia743

ICA_Normal, individual class accuracy for Normal class; ICA_Pneumonia, individual class accuracy for Pneumonia class.

From the results of Experiment 6, as shown in Table 6.1, it can be seen that the hybrid CAC system designed using the deep features extracted from the GoogLeNet CNN model fed to PCA-SVM classifier achieves 91.00% accuracy for the classification of chest radiograph images into binary classes: Normal and Pneumonia. The individual class accuracy value of Normal class is 96.00%, and for the Pneumonia class, the individual class accuracy value obtained is 86.00%. From the total 100 images in the testing set, 9 images have been incorrectly classified, from which 2 images belong to the Normal class and 7 images belong to the Pneumonia class. The ROC curve with its corresponding AUC values for the hybrid CAC system designed using the GoogLeNet CNN model as deep feature extractor and PCA-SVM classifier is shown in Fig. 6.4.

Fig. 6.4
Fig. 6.4 The ROC curve with its corresponding AUC values for the hybrid CAC system designed using the GoogLeNet CNN model as deep feature extractor and PCA-SVM classifier.

6.7: Concluding remarks

This chapter gives a detailed overview of the PCA feature dimensionality technique and explains in detail the steps involved for classification of the images using the SVM classifier. From the experiment carried out in this chapter, it is observed that designing the hybrid CAC system through deep feature extraction by GoogLeNet and PCA-SVM achieves 91.00% accuracy for the classification of chest radiographs. The next chapter aims at understanding the lightweight CNN model, its architecture, and experiments conducted to evaluate the performance of designing lightweight CNN-based CAC systems for classification of chest radiographs.

References

[1] Jain I., Jain V.K., Jain R. Correlation feature selection based improved-binary particle swarm optimization for gene selection and cancer classification. Appl. Soft Comput. 2018;62:203–215.

[2] Michalak K., Kwaśnicka H. Correlation-based feature selection strategy in classification problems. Int. J. Appl. Math. Comput. Sci. 2006;16:503–511.

[3] Hall M.A. Correlation-based feature selection for discrete and numeric class machine learning. In: Proceedings of the Seventeenth International Conference on Machine Learning, Morgan Kaufmann Publishers Inc.; 2000:359–366.

[4] Toğaçar M., Ergen B., Cömert Z., Özyurt F. A deep feature learning model for pneumonia detection applying a combination of mRMR feature selection and machine learning models. IRBM. 2020;41(4):212–222.

[5] Liu Q., Gu Q., Wu Z. Feature selection method based on support vector machine and shape analysis for high-throughput medical data. Comput. Biol. Med. 2017;91:103–111.

[6] Hsu H.H., Hsieh C.W. Feature selection via correlation coefficient clustering. JSW. 2010;5(12):1371–1377.

[7] Chandrashekar G., Sahin F. A survey on feature selection methods. Comput. Electr. Eng. 2014;40(1):16–28.

[8] Jović A., Brkić K., Bogunović N. A review of feature selection methods with applications. In: 2015 38th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO). IEEE; 2015:1200–1205.

[9] Hua J., Tembe W.D., Dougherty E.R. Performance of feature-selection methods in the classification of high-dimension data. Pattern Recogn. 2009;42(3):409–424.

[10] Remeseiro B., Bolon-Canedo V. A review of feature selection methods in medical applications. Comput. Biol. Med. 2019;112:103375.

[11] Allam M., Nandhini M. A study on optimization techniques in feature selection for medical image analysis. Int. J. Comput. Sci. Eng. 2017;9(3):75–82.

[12] Tang J., Alelyani S., Liu H. Feature selection for classification: a review. In: Data Classification: Algorithms and Applications. CRC Press; 2014:37–64. https://doi.org/10.1201/b17320.

[13] Jolliffe I.T., Cadima J. Principal component analysis: a review and recent developments. Philos. Trans. Royal Soc. A Math. Phys. Eng. Sci. 2016;374(2065):20150202.

[14] Brems M. A one-stop shop for principal component analysis. Medium Towards Data Science. 2017;17.

[15] Powell V., Lehe L. Principal Component Analysis Explained Visually. DISQUS, Available: http://setosa.io/ev/principal-componentanalysis/. 2015.

[16] Rasheed J., Hameed A.A., Djeddi C. A machine learning-based framework for diagnosis of COVID-19 from chest X-ray images. Interdiscip. Sci. Comput. Life Sci. 2021;13:103–117. doi:10.1007/s12539-020-00403-6.

[17] Wang L., ed. Support Vector Machines: Theory and Applications. Springer Science & Business Media; . 2005;vol. 177.

[18] Pradhan A. Support vector machine—a survey. Int. J. Emerging Technol. Adv. Eng. 2012;2(8):82–85.

[19] Suthaharan S. Support vector machine. In: Machine Learning Models and Algorithms for Big Data Classification. Boston, MA: Springer; 2016:207–235.

[20] Zhou J., Chan K.L., Chong V.F.H., Krishnan S.M. Extraction of brain tumor from MR images using one-class support vector machine. In: 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference. IEEE; 2006:6411–6414.

[21] Jiang Y., Li Z., Zhang L., Sun P. An improved svm classifier for medical image classification. In: International Conference on Rough Sets and Intelligent Systems Paradigms. Berlin, Heidelberg: Springer; 2007:764–773.

[22] Chang C.C., Lin C.J. LIBSVM: A Library of Support Vector Machines. Software available at http://www. csie. ntu. edu. tw/~ cjlin/libsvm. 2012.

[23] Sánchez A.V.D. Advanced support vector machines and kernel methods. Neurocomputing. 2003;55(1–2):5–20.

[24] Karatzoglou A., Meyer D., Hornik K. Support vector machines in R. J. Stat. Softw. 2006;15(9):1–28.

[25] Pelckmans K., Suykens J.A., Van Gestel T., De Brabanter J., Lukas L., Hamers B., De Moor B., Vandewalle J. LS-SVMlab: a matlab/c toolbox for least squares support vector machines. 2002 Tutorial. KULeuven-ESAT. Leuven, Belgium, 142 (1–2).

[26] Meyer D., Wien F.T. Support vector machines. The Interface to libsvm in package e 1071, 28. 2015.

[27] Gomes T.A., Prudêncio R.B., Soares C., Rossi A.L., Carvalho A. Combining meta-learning and search techniques to select parameters for support vector machines. Neurocomputing. 2012;75(1):3–13.

[28] Virmani J., Kumar V., Kalra N., Khandelwal N. SVM-based characterization of liver ultrasound images using wavelet packet texture descriptors. J. Digit. Imaging. 2013;26(3):530–543.

[29] Virmani J., Dey N., Kumar V. PCA-PNN and PCA-SVM based CAD systems for breast density classification. In: Applications of Intelligent Optimization in Biology and Medicine. Cham: Springer; 2016:159–180.

[30] Virmani J., Kumar V., Kalra N., Khandelwa N. PCA-SVM based CAD system for focal liver lesions using B-mode ultrasound images. Def. Sci. J. 2013;63(5):478–486.

[31] Virmani J., Kumar V., Kalra N., Khandelwal N. SVM-based characterisation of liver cirrhosis by singular value decomposition of GLCM matrix. Int. J. Artif. Intell. Soft Comput. 2013;3(3):276–296.

[32] Rana S., Jain S., Virmani J. SVM-based characterization of focal Kidney lesions from B-mode ultrasound images. JUIT. (0975-8585):2016. http://ir.juit.ac.in/123456789/6908.

[33] Hassanein A.E., Kim T.H. Breast cancer MRI diagnosis approach using support vector machine and pulse coupled neural networks. J. Appl. Logic. 2012;10(4):274–284.

[34] Chan Y.H., Zeng Y.Z., Wu H.C., Wu M.C., Sun H.M. Effective pneumothorax detection for chest X-ray images using local binary pattern and support vector machine. J. Healthc. Eng. 2018;2018:2908517.

[35] Hassanien A.E., Mahdy L.N., Ezzat K.A., Elmousalami H.H., Ella H.A. Automatic x-ray covid-19 lung image classification system based on multi-level thresholding and support vector machine. med Rxiv. 2020.

[36] Sousa R.T., Marques O., Soares F.A.A., Sene Jr. I.I., de Oliveira L.L., Spoto E.S. Comparative performance analysis of machine learning classifiers in detection of childhood pneumonia using chest radiographs. Procedia Comput. Sci. 2013;18:2579–2582.

[37] Depeursinge A., Iavindrasana J., Hidki A., Cohen G., Geissbuhler A., Platon A., Poletti P.A., Müller H. Comparative performance analysis of state-of-the-art classification algorithms applied to lung tissue categorization. J. Digit. Imaging. 2010;23(1):18–30.

[38] Yao J., Dwyer A., Summers R.M., Mollura D.J. Computer-aided diagnosis of pulmonary infections using texture analysis and support vector machine classification. Acad. Radiol. 2011;18(3):306–314.

[39] Naydenova E., Tsanas A., Casals-Pascual C., De Vos M. Smart diagnostic algorithms for automated detection of childhood pneumonia in resource-constrained settings. In: 2015 IEEE Global Humanitarian Technology Conference (GHTC). IEEE; 2015:377–384.

Further reading

[40] Nixon M., Aguado A. Feature Extraction and Image Processing for Computer Vision. Academic Press; 2019.

[41] Yang A., Yang X., Wu W., Liu H., Zhuansun Y. Research on feature extraction of tumor image based on convolutional neural network. IEEE Access. 2019;7:24204–24213.

[42] Srinivas M., Roy D., Mohan C.K. Discriminative feature extraction from X-ray images using deep convolutional neural networks. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2016:917–921.

[43] Chouhan V., Singh S.K., Khamparia A., Gupta D., Tiwari P., Moreira C., Damaševičius R., De Albuquerque V.H.C. A novel transfer learning based approach for pneumonia detection in chest X-ray images. Appl. Sci.. 2020;10(2):559.

[44] Ravishankar H., Sudhakar P., Venkataramani R., Thiruvenkadam S., Annangi P., Babu N., Vaidya V. Understanding the mechanisms of deep transfer learning for medical images. In: Deep Learning and Data Labeling for Medical Applications. Cham: Springer; 2016:188–196.

[45] Wu H., Xie P., Zhang H., Li D., Cheng M. Predict pneumonia with chest X-ray images based on convolutional deep neural learning networks. J. Intell. Fuzzy Syst.. 2020;1–15 (Preprint).

[46] Suzuki K. Overview of deep learning in medical imaging. Radiol. Phys. Technol.. 2017;10(3):257–273.

[47] Wibisono A., Adibah J., Priatmadji F.S., Viderisa N.Z., Husna A., Mursanto P. Segmentation-based knowledge extraction from chest X-ray images. In: 2019 4th Asia-Pacific Conference on Intelligent Robot Systems (ACIRS). IEEE; 2019:225–230.

[48] Anwar S.M., Majid M., Qayyum A., Awais M., Alnowami M., Khan M.K. Medical image analysis using convolutional neural networks: a review. J. Med. Syst.. 2018;42(11):226.

[49] Dey N., Ashour A.S., Shi F., Balas V.E. Soft Computing Based Medical Image Analysis. Academic Press; 2018.

[50] Kumar I., Virmani J., Bhadauria H.S., Panda M.K. Classification of breast density patterns using PNN, NFC, and SVM classifiers. In: Soft Computing Based Medical Image Analysis. Academic Press; 2018:223–243.

[51] Badnjevic A., Gurbeta L., Custovic E. An expert diagnostic system to automatically identify asthma and chronic obstructive pulmonary disease in clinical settings. Sci. Rep.. 2018;8(1):1–9.

[52] García-Floriano A., Ferreira-Santiago Á., Camacho-Nieto O., Yáñez-Márquez C. A machine learning approach to medical image classification: detecting age-related macular degeneration in fundus images. Comput. Electr. Eng.. 2019;75:218–229.

[53] Oliveira L.L.G., e Silva S.A., Ribeiro L.H.V., de Oliveira R.M., Coelho C.J., Andrade A.L.S. Computer-aided diagnosis in chest radiography for detection of childhood pneumonia. Int. J. Med. Inform.. 2008;77(8):555–564.

[54] Simon P., Uma V. Deep learning based feature extraction for texture classification. Procedia Comput. Sci.. 2020;171:1680–1687.

[55] Boyd A., Czajka A., Bowyer K. Deep learning-based feature extraction in iris recognition: use existing models, fine-tune or train from scratch?. In: 2019 IEEE 10th International Conference on Biometrics Theory, Applications and Systems (BTAS). IEEE; 2019:1–9.

[56] O’Mahony N., Campbell S., Carvalho A., Harapanahalli S., Hernandez G.V., Krpalkova L., Riordan D., Walsh J. Deep learning vs. traditional computer vision. In: Science and Information Conference. Cham: Springer; 2019:128–144.

[57] Dara S., Tumma P. Feature extraction by using deep learning: a survey. In: 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA). IEEE; 2018:1795–1801.

[58] Dey N., Zhang Y.D., Rajinikanth V., Pugalenthi R., Raja N.S.M. Customized VGG19 architecture for pneumonia detection in chest X-rays. Pattern Recogn. Lett.. 2021;143:67–74.

[59] Varshni D., Thakral K., Agarwal L., Nijhawan R., Mittal A. Pneumonia detection using CNN based feature extraction. In: In 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT). IEEE; 2019:1–7.

[60] Hashmi M.F., Katiyar S., Keskar A.G., Bokde N.D., Geem Z.W. Efficient pneumonia detection in chest xray images using deep transfer learning. Diagnostics. 2020;10(6):417.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset