Human-centric Computing and Information Sciences volume 12, Article number: 24 (2022)
Cite this article 1 Accesses
https://doi.org/10.22967/HCIS.2022.12.024
Since chest illnesses are so frequent these days, it is critical to identify and diagnose them effectively. As such, this study proposes a model designed to accurately predict chest disorders by analyzing multiple chest x-ray pictures obtained from a dataset, consisting of 112,120 chest X-ray images, obtained the National Institute of Health (NIH) X-ray. The study used photos from 30,805 individuals with a total of 14 different types of chest disorder, including atelectasis, consolidation, infiltration, and pneumothorax, as well as a class called “No findings” for cases in which the ailment was undiagnosed. Six distinct transfer-learning approaches, namely, VGG-16, MobileNet V2, ResNet-50, DenseNet-161, Inception V3, and VGG-19, were used in the deep learning and federated learning environment to predict the accuracy rate of detecting chest disorders. The VGG-16 model showed the best accuracy at 0.81, with a recall rate of 0.90. As a result, the F1 score of VGG-16 is 0.85, which was higher than the F1 scores computed by other transfer learning approaches. VGG-19 obtained a maximum rate of accuracy of 97.71% via federated transfer learning. According to the classification report, the VGG-16 model is the best transfer-learning model for correctly detecting chest illness.
Deep Learning, Chest Diseases, Federated Learning, Disease Prediction, X-Ray Dataset
The chest or thorax is situated between the neck and abdomen, and is one of the three main parts of a human body. It comprises the heart, lungs, muscles, and many other regional compositions. Some diseases or infections affect the chest area, such as atelectasis, consolidation, and pleural thickening among others. X-rays of the chest are mainly taken to detect such infections [1]. Such X-rays may show cavities, filtrates, nodules and the like, thereby helping to diagnose chest disease. Chest diseases such as pneumonia, asthma, and lung diseases are serious health disorders that can have grave negative effects on human health. Detecting a chest infection is a laborious task; consequently, researchers have proposed various kinds of systems for this purpose [2], and many techniques and algorithms have been applied to detect diverse types of chest infections. Previously, machine learning (ML) was widely in the medical field to detect health diseases. ML techniques have a significant classification characteristic classifying new cases or observations based on previous ones. The type of disease and its severity are measured using various ML techniques such as naive Bayes, decision trees, k-nearest neighbor, and support vector machine (SVM) among others [3]. Deep learning represents the next major advance in the detection of diseases. A deep learning subset of ML can be used to train artificial intelligence machines to predict outputs and extract data patterns using an artificial neural network. It is being applied in diverse fields such as defense, security, voice recognition, face recognition, disease detection, and so forth [4].
Deep learning focuses mainly on several modifications, feature selection, and resizing made to pre-process the inputs [5, 6]. Deep learning is very popular in medical diagnosis as it extracts valuable characteristics from input images [7]. Another technique explored in the medical diagnosis of diseases is transfer learning or fine-tuning. It is also a type of machine learning that reuses a model previously developed for certain tasks. The reusable model is already pre-trained to classify certain features that help to predict the disease easily and quickly. These models need not be trained again, as they are already fine-tuned for medical applications. Their use in the medical field consists in detecting planes in ultrasounds, and classifying lung diseases and types of chest infections [8,9]. Federated learning is a theory presented by Google to construct models based on ML that could work on multiple devices adhering to the privacy of data [10, 11]. It is an up-graded form of artificial intelligence based on the central idea of ensuring users’ data privacy. A globally shared model is used in the device so that it trained collectively, thus providing proper security. As stated in [12], there are three types of federated learning, as follows: horizontal or sample-based federated learning in which all devices have the same features of data; vertical federated learning, which involves different data comprising varying features to train a model collectively; and federated transfer learning (FTL), which solves a new problem by utilizing the transfer learning features obtained by applying an already trained model for another task. On the other hand, wearable gadgets have grown in popularity recently, with a wide range of applications in health monitoring systems, resulting in the growth of the “Internet of Medical Things” (IoMT). The IoMT has a critical role to play in lowering death rates by detecting chest illnesses early [13, 14]. It comprises various items of medical equipment connected to a healthcare provider’s computer system via the Internet, which are capable of producing, storing, analyzing, and disseminating health data [15]. Wearables, remote patient monitoring, sensor-enabled beds, infusion pumps, and health tracking devices are all IoMT items. The purpose of the IoMT is to improve the satisfaction of both patients and healthcare providers, and to ensure the quality of treatment.
The model proposed in this study employs deep and FTL techniques to detect chest diseases through X-rays images. The first step performed by the system is data pre-preprocessing, which involves loading and cleaning data and converting them into a more intelligible form. This step is followed by exploratory data analysis, which is used to categorize and study data based on various attributes, such as gender, age, single or multiple diseases. Furthermore, several features such as width, area, perimeter, epsilon, height, solidity, etc., are extracted for each disease. Training and testing of the data are followed by data augmentation in order to make the images more transparent. Finally, various deep and federated transfer-learning models, such as VGG-16, MobileNet V2, ResNet-50, DenseNet-161, Inception V3, and VGG-19, make disease predictions. Keras plus TensorFlow is used for deep learning, while federated learning is employed on PyTorch technology. For the analysis conducted in this study, a dataset named “National Institutes of Health (NIH) [16] X-ray dataset” and comprising 112,120 images of various chest infections was used. In addition, there are disease labels from approximately 30,805 patients and a total of 14 classes of diseases such as atelectasis, consolidation, infiltration, pneumothorax, etc., and a class named “No findings” for cases where a disease remained undetected. The rest of the paper is arranged as follows: Section 2 discusses previous literature in the relevant research field; Section 3 describes the dataset, techniques, and libraries used; Section 4 presents the applied models and the results; and, finally, Section 5 discusses the future scope of the work.
Several recent studies focused on the identification of various forms of chest infections. All such works employ a distinctive methodology and are increasingly reliant on other learning strategies. Some of them are included in the section on related works. Deep learning is a type of machine learning technology designed to mimic the behavior of the human brain. Because they are constantly evaluating data, these algorithms attempt to emulate and consider aspects of human behavior [6]. Deep learning algorithms make use of a multi-layered architecture composed of several levels. Recently, there has been a significant increase in the use of deep learning to forecast many types of infections and disorders. Deep learning combined with big data is being used to forecast infectious illnesses, according to a study proposal by Chae et al. [17]. The deep neural network (DNN), long short-term memory (LSTM) models, autoregressive integrated moving average (ARIMA), and ordinary least squares (OLS) approach are used on a variety of various forms of data, including weather data, twitter data, and non-clinical search data, in order to forecast illnesses. The results showed that the DNN models were superior to the others in terms of average best performance, but the LSTM model was more accurate in predicting when an infectious illness was spreading. The DNN and LSTM models also outperformed the ARIMA model in terms of accuracy. Abuhamdah et al. [3] presented a hybrid predictive model comprising a convolution neural network using WekaDeepLearning. The hybrid model was used to detect pneumonia and other lung diseases using chest X-ray images or computed tomography (CT) images. In a nutshell, their study showed the capability and effectiveness of the classifier in detecting positive cases through an experimental performance. Kishor et al. [18] worked on improving the quality of service provided by a heterogeneous network using a “reinforcement learning-based multimedia data segregation” (RLMDS) algorithm and a “computing QoS in medical information system using fuzzy” (CQMISF) algorithm in fog computing. The authors’ main aim was to use the proposed algorithms to classify data and transfer the classified high-risk data to the end-users by selecting the optimal gateway. Bhattacharyya et al. [19] presented a study aimed at classifying chest-related diseases such as pneumonia and COVID-19 (coronavirus disease 2019) with normal X-ray images. The authors used a conditional generative adversarial network to segment the images and deep learning models in order to extract discriminatory features. Convolutional neural network (CNN), Back-propagation neural networks (BPNNs) with supervised learning, and competitive neural networks with unsupervised learning are all types of neural networks used to diagnose pulmonary disorders [20]. It has been demonstrated that CNN can perform classification better than other models thanks to its extensive deep structure, which can extract features at various levels of abstraction and complexity. In addition, this model has a high rate of identification of its features.
Transfer learning is another technology that is becoming increasingly popular in the diagnosis of illnesses [21]. When it comes to transferring learning, the theory behind it holds that a model previously established for a specific job may be used as a starting point. To give an example, the study by Hon and Khan [10] resulted in the development of a categorization system for Alzheimer’s disease. Using pre-trained weights, models such as Inception V4 and VGG-16 were deployed on big datasets, and the image entropy method was used to choose the most informative images for training purposes. It was intended to alleviate the limitation of having to train existing algorithms on many photos. VGG-16 with transfer learning and Inception V4 produced accuracy rates of 92.3% and 96.25%, respectively, demonstrating that just a small number of training pictures is required to obtain correct results in this study. Muniasammy et al. [9] proposed a deep learning-based model to diagnose chest disease information provided in the form of images as well as medical reports. The main purpose of this study was that that model should be able to automatically detect chest diseases from various chest X-ray-based images using various class labels. Kishor et al. [22] used six ML algorithms, namely, decision tree, SVM, naive Bayes, random forest, artificial neural network, and k-nearest neighbor, to detect nine severe diseases including heart disease, diabetics, breast cancer, hepatitis, liver disorder, dermatology, surgery data, etc. While applying all these algorithms, the random forest classifier observed a maximum accuracy rate of 97.62%, sensitivity of 99.67%, specificity of 97.81%, and AUC (area under the curve) of 99.32% for different diseases. In the study by Vogado et al. [23], transfer learning was utilized in CNNs and SVMs to create a system for diagnosing leukemia. With this method, an input picture is sent to the CNN for feature extraction; the gain ratio is utilized to pick features; and, lastly, the SVM is employed as a classifier for classification. There is a significant distinction between the suggested technique and current state-of-the-art methods in that the input photos are used directly without any pre-processing being performed. Therefore, it is not necessary to go through the process of segmentation. Chest infections can manifest themselves in a variety of ways, including consolidation, pleural thickening, and cardiomegaly among others. Table 1 shows a list of some extant works that are relevant to these topics [24–36]. Deep transfer learning (DTL) has proven to be extremely effective in detecting a wide range of illnesses and disorders. Pathak et al. [37] presented a DTL-based approach to categorizing COVID-19, among other illnesses. A chest CT dataset is inputted into a ResNet-50 network, and is then used to perform feature extraction on the dataset. To forecast whether COVID-19 is positive or negative in the input sample, transfer learning uses these characteristics, which are used as parameters in deep CNN. The model’s testing accuracy is 93.01%, making it a viable alternative to a COVID-19 testing kit in some situations. A DTL-based model for predicting COVID-19 in chest X-ray images was also developed by Minaee et al. [38] in a similar fashion. Transfer learning is used to train four CNN models, namely, ResNet-18, ResNet-50, SqueezeNet, and DenseNet-121, of which ResNet-18 is the most widely used CNN model. Data augmentation techniques such as flipping, rotation, and other image manipulation techniques change pictures to increase the number of samples. Fine-tuning is performed on the last layer of the pre-trained model using ImageNet. According to the data, the sensitivity rate is 98%, and the average specificity rate is 90%. A deep CNN with transfer learning was also proposed by Rahman et al. [39] to detect pneumonia. Four pre-trained deep CNN models, i.e.,AlexNet, ResNet-18, DenseNet-201, and SqueezeNet, were used in this study as follows: initially, the system takes X-ray pictures from the X-ray machine and stores them, after which they are sent to the ML block, which pre-processes (resizing and normalization) and augments the data (rotation, scaling, translation). The output of the pre-trained models is characterized as follows: regular pneumonia, bacterial pneumonia, or viral pneumonia. DenseNet-201 exceeds all other deep CNN networks in terms of performance. Sahaand Rahman [5] applied the convolution neural network method to predict the presence of pneumonia using chest X-ray images. Their model showed a rate of accuracy of 89%, which was better than the existing deep learning-based clinical image classification algorithms.
In the survey of previous papers, it was observed that the categorization of chest-based illness was performed using limited or restricted datasets. However, in this study a large dataset, composed of 112,120 [16] chest X-ray images, was obtained in order to classify each chest related disease.
Comparing with state-of-the-art research efforts, this study integrated various pre-transfer learning models with a deep and federated learning mechanism in order to make a fair comparison and obtain better performance. Hence, for the fourteen distinct types of chest X-ray disorders, classifications were done using the following six transfer learning models: VGG-16, MobileNetV2, ResNet-50, DenseNet-161, Inception V3, and VGG-19.
Study | Chest infection type | Dataset | Approach used | Results |
---|---|---|---|---|
Liu et al. [24] | Atelectasis | 130 patients of Beijing Military General Hospital | Deep neural network | Sensitivity: |
Ultrasound = 100% | ||||
Chest X-ray = 75% | ||||
Ullmann et al. [25] | Atelectasis | 40 children affected by neuromuscular disease | Neural network | LUS: |
Specificity = 82% | ||||
Sensitivity = 57% | ||||
Positive predictive value = 80% | ||||
Negative predictive value = 61% | ||||
Behzadi-Khormouji et al. [26] | Consolidation | Pediatric chest X-ray dataset and ImageNet | VGG-16, DenseNet-121, ChestNet, PyramidCNN | Accuracy = 94.67% |
The ChestNet2 model outperformed the other five models. | ||||
Na'am et al. [27] | Infiltration | X-ray images of infants treated at Central Public Hospital (RSUP), Indonesia | Morphological operations, edge detection, and sharpening of edges. | Output images showed clearer edges and easy recognizable information. |
Gooßen et al. [28] | Pneumothorax | 1,003 chest X-ray images | CNN, multiple-instance learning (MIL), fully convolutional networks (FCN). | AUC: |
CNN = 0.96 | ||||
MIL = 0.93 | ||||
FCN = 0.92 | ||||
Chan et al. [29] | Pneumothorax | 32 pneumothorax and 10 normal chest radiographs from Chung Shan Medical University Hospital, Taiwan | Support vector machines (SVMs) | Average accuracy = 82.2% |
Local binary pattern (LBP) | ||||
Campo et al. [30] | Emphysema | 7,377 images | 11-layer CNN using percentage of low-attention lung areas (LAA, %) | Mean error = 3.96 |
AUC accuracy = 90.73% | ||||
Mean sensitivity = 85.68% | ||||
Jain et al. [31] | Pneumonia | Chest X-ray images dataset: 5,216 training and 624 testing images | Six models: first and second model consisting of two and three convolutional layers, respectively, VGG-16, VGG-19, ResNet-50, and Inception V3. | Validation accuracy of first model (85.26%), second model (92.31%) |
Accuracy of other four models: | ||||
VGG-16 = 87.28% | ||||
VGG-19 = 88.46% | ||||
ResNet-50 = 77.56% | ||||
Inception-V3 = 70.99%, | ||||
Hashmi et al. [32] | Pneumonia | 700 testing set images, pneumonia dataset | ResNet-18, DenseNet-121, Inception V3 | Validation accuracy = 98.43%. |
Saito et al. [33] | Pleural thickening | 28,727 chest X-rays of students and employees at the University of Tokyo | Two-tailed Student t-tests, the chi-square test, and binary logistic regression | More than 90% of the cases were defined as pulmonary apical cap. Frequency increased with age; more prevalent in males and smokers. Also, persons with low body weight along with tall height were more prone to this. |
Alghamdi et al. [34] | Cardiomegaly | 59 patients, Abdul-Aziz Hospital, Jeddah, Saudi Arabia | Cardio-thoracic ratio (CTR) | 21 patients with cardiomegaly; patients aged 37–58 were the most affected; more prevalent among males. |
Candemir et al. [35] | Cardiomegaly | NLM-Indiana Collection and NIH-CXR dataset | Pre-trained models are used: fine-tuning and CXR-based. | Accuracy = 89.86% |
Sensitivity = 88.81% | ||||
Specificity = 90.91% | ||||
Liang et al. [36] | Nodulo mass | 100 patients, Kaohsiung Veterans General Hospital, Kaohsiung, Taiwan | Heat map, abnormal probability, nodule probability, mass probability. | Detection performance with 76.6% sensitivity and 88.68% specificity. |
This work uses a novel methodology in which a transfer-learning model is combined with deep and federated learning to classify chest diseases. Initially, the chest data were collected in the form of images from NIH chest X-rays, which were then pre-processed to clean the data, match then with the .csv dataset, find the NAN values, and encode the data. Further, the preprocessed data were visualized and summarized graphically to assist with feature extraction and to extract contour features such as area, perimeter, aspect ratio, solidity, etc.
Later data were split into training and testing datasets (75% and 25%, respectively), which were then further augmented by various techniques such as flipping, rotation, etc. Finally, pre-trained models such as VGG-16, MobileNet V2, ResNet-50, DenseNet-161, Inception V3, and VGG-19 were used to perform classification, which was further evaluated using the precision rate, recall rate, and AUC & F1 score, as shown in Fig. 1.
The proposed system utilizes transfer learning to apply already pre-trained models, helping to reduce the training time. These models can be used as an approach to solving a new problem. A pre-trained model can be directly embedded into an application or used as a model to extract features for classification. Several models perform well in the image classification process, and some are listed in the following section.
$Accuracy= (TP+TN)/ (TP+FP+TN+FN)$(1)
Loss: The difference between the ground truth and the predicted value is defined as Loss.$RMSE= \sqrt{\frac{∑_{i=1}^N (x_i-\hat x_i)^2}{N}}$(2)
Here $i$ is a variable, $N$ is non missing data points, $x_i$ is actual observations of time series, and $\hat x_i$ is estimated time series.$Precision= TP/(TP+FP)$(3)
Recall: It can be defined as the ratio of positive instances that are accurately predicted. Mathematically, it is calculated by using Equation (4):$Recall= TP/(TP+TN)$(4)
F1 score: It is the weighted ratio of precision and recall. A high F1 score indicates that the model has a good classifying ability [52]. It can be calculated by using Equation (5):$F1= (2×precision×recall)/(precision+recall)$(5)
AUC: Area under curve is defined as the definite integral of the curve that describes the variation of expressions in a human being as a function of time. Its mathematical formula is shown in Equation (6):$A= \lim\limits_{x→∞} \displaystyle\sum_{i=1}^n f(x)$(6)
Here, TP stands for true positive, TN stands for true negative, FP stands for false positive, and FN stands for false negative.Model | Training | Validation | ||||
---|---|---|---|---|---|---|
Accuracy (%) | Loss | RMSE | Accuracy (%) | Loss | RMSE | |
VGG-16 | 87.33 | 0.447 | 0.668 | 86.43 | 0.72 | 0.848 |
MobileNet V2 | 85.46 | 0.468 | 0.684 | 88.09 | 0.582 | 0.762 |
ResNet-50 | 87.46 | 0.41 | 0.64 | 87.41 | 0.355 | 0.595 |
DenseNet-161 | 88.01 | 0.561 | 0.748 | 88 | 0.456 | 0.675 |
Inception V3 | 87.98 | 0.324 | 0.569 | 87.94 | 0.321 | 0.566 |
Model | Training | Validation | ||||
---|---|---|---|---|---|---|
Accuracy (%) | Loss | RMSE | Accuracy (%) | Loss | RMSE | |
VGG-16 | 78.65 | 0.127 | 0.356 | 76.71 | 0.213 | 0.461 |
MobileNet V2 | 79.23 | 0.126 | 0.355 | 79 | 0.194 | 0.44 |
ResNet-50 | 79.29 | 0.186 | 0.431 | 79.07 | 0.194 | 0.44 |
DenseNet-161 | 80.02 | 0.186 | 0.431 | 82.71 | 0.193 | 0.44 |
Inception V3 | 94.56 | 0.187 | 0.432 | 94.9 | 0.185 | 0.431 |
Model | Precision | Recall | F1 score |
---|---|---|---|
VGG-16 |
0.81 | 0.90 | 0.85 |
MobileNet V2 |
0.79 | 0.89 | 0.84 |
ResNet-50 |
0.76 | 0.73 | 0.76 |
DenseNet-161 |
0.80 | 0.78 | 0.84 |
Inception V3 |
0.78 | 0.88 | 0.82 |
VGG-19 |
0.78 | 0.88 | 0.83 |
Algorithm | Computation time (s) |
---|---|
VGG-16 | 37,831 |
VGG-19 | 36,910 |
MobileNet V2 | 27,958 |
ResNet-50 | 29,150 |
DenseNet-161 | 24,645 |
Inception V3 | 35,284 |
Model | Image | Predicted disease | |
---|---|---|---|
Deep learning | Federated learning | ||
VGG-16 | Fig. 5(a) | Atelectasis | Atelectasis |
VGG-19 | Fig. 5(b) | Effusion | Effusion |
MobileNet V2 | Fig. 5(c) | Mass | Mass |
ResNet-50 | Fig. 5(d) | Pneumothorax | Pneumothorax |
DenseNet-161 | Fig. 5(e) | Infiltration | Infiltration |
Inception V3 | Fig. 5(f) | Mass, Infiltration | Mass, Infiltration |
Technique | Accuracy (%) |
---|---|
Pre-trained models [35] | 89.86 |
Deep learning-based decision tree classifier [53] | 89 |
CNN [7] | 91.24 |
Proposed model (VGG-19) | 97.7 |
Technique | Dataset | Accuracy (%) |
---|---|---|
CNN [54] | Chest Xray | 98.9 |
ResNet [55] | ImageNet | 96.1 |
VGG-16 [56] | Chest Xray | 87 |
Proposed model (VGG-19) | NIH dataset | 97.7 |
Detecting chest disorders via chest X-rays is a complex and essential undertaking in the field of human health. Although much work has already been done on this subject, employing deep and federated learning approaches combined with transfer learning models to identify chest ailments is a novel approach to this task. Using pre-trained models makes detection easier and enhances the system’s accuracy and efficiency. As a result, this work has effectively detected chest ailments by using an integrated transfer learning strategy combined with deep and federated learning approaches. The proposed model used the NIH’s chest illness X-ray dataset, which includes 112,120 chest pictures and the names of the conditions of 30,805 persons. There are 14 different types of chest illnesses. The system first performed pre-processing and exploratory data analysis on the input photos to extract useful information, which was then utilized to extract features rapidly and accurately. The data for training and testing were then separated in order to perform data augmentations. Finally, the data were subjected to various transfer learning models, including VGG-16, MobileNet V2, ResNet-50, DenseNet-161, Inception V3, and VGG-19, the results of which were assessed to determine which model is most suited to identify chest disorders. MobileNet V2 (88.09%) was shown to outperform the other deep learning models in terms of validation accuracy. In contrast, the VGG-19 model outperformed the other transfer learning models in terms of validation accuracy when used with federated learning, with a validation accuracy of 97.71%. Overall, the VGG-16 model proved to have the best accuracy and recall rate. As a result, the F1 score of VGG-16 is automatically the highest. As a result, the classification report reveals that the VGG-16 model detects chest infections accurately and effectively, making it the best transfer learning model for this purpose. With these discoveries, our technique for integration with any system will allow pulmonologists and radiologists to correctly diagnose chest diseases by utilizing various medical imaging sources in less time with better flexibility. However, this study has a drawback in that the proposed approach was only able to categorize the fourteen illnesses included in the dataset.
Moreover, to achieve the best outcomes, the models should also integrate optimization strategies, and the execution time should be factored in. In reality, doctors should aim to incorporate such technologies into their regular diagnosing of patients’ diseases. Furthermore, by recognizing chest problems early on and providing proper treatment, the proposed model could save countless lives. Further classifications of chest problems could be included in the system to produce better results in detecting any type of chest ailment.
Jana Shafi would like to thank the Deanship of Scientific Research, Prince Sattam Bin Abdul Aziz University, for supporting this work.
Conceptualization, MB. Funding acquisition, YK, YS, JS. Investigation and methodology, BK, PJ, YK, HP. Project administration, YK, YS, JS. Resources, PJ, YK, YS, JS. Supervision, PJ, YK, YS. Writing of the original draft, BK, PJ, YK, HP. Writing of the review and editing, HP, YS, JS. Software, BK, PJ. Validation, YK, HP, YS, JS. Formal analysis, BK, PJ, YK. Visualization, HP, YS, JS.
This work was supported by the National Research Foundation of Korea (NRF), funded by the Ministry of Science and ICT (MSIT) of the Korean government (Grant No. 2020R1C1C1003425 and 2020R1A4A3079710).
The authors declare that they have no competing interests.
Name : Barkha Kakkar
Affiliation : Galgotia’s University,India
Biography : Barkha Kakkar is pursuing PhD in Computer Application from, Galgotia’s University, Greater Noida, India. She is also working as Assistant Professor in Institute of Technology & Science, Mohan Nagar Ghaziabad She completed his M.C.A. from AKGEC Affiliated to Uttar Pradesh Technical University Lucknow. Her research area is Blockchain in Healthcare. 2022.)
Name : Prashant Johri
Affiliation : Galgotia’s University,India
Biography : Prashant Johri is currently a, Professor in School of Computing Science & Engineering, Galgotias University, Greater Noida, India. He received his B.Sc.(H) , M.C.A. from A.M.U. and Ph.D. in Computer Science from Jiwaji University, Gwalior in 2011, India. He has also worked as a Professor and Director (M.C.A.) in G.I.M.T. and N.I.E.T. Gr.Noida. He has supervised 2 PhD students and M.Tech. Students for their thesis. He has published 150 scientific articles. He has published edited books. His research interest includes Artificial Intelligence, Machine Learning, Data Science, Cloud Computing, Block Chain,Healthcare, Agriculture, Image Processing, Software Reliability.
Name : Yogesh Kumar
Affiliation : Indus University,India
Biography : Yogesh Kumar is working as Associate Professor at Indus Institute of Technology & Engineering, Indus University, Rancharda, Ahmedabad. He has done his Ph.D. CSE from Punjabi University, Patiala. Prior to this, he has done his M.Tech CSE from Punjabi university, Patiala. He is having total 14 years of experience including teaching and research with more than 57 Publications in various reputed journals. His research areas include Artificial Intelligence, Deep Learning and Computer Vision.
Name : Hyunwoo Park
Affiliation : Dongguk University-Seoul, Korea
Biography : Hyunwoo Park was born in Cheongju, Republic of Korea in 1996. He received the B.S. degree in Industrial and Systems Engineering from Dongguk University-Seoul, Seoul, Republic of Korea, in 2020. He is currently pursuing the M.S. degree in Industrial and Systems Engineering at the Dongguk University-Seoul. His research interests include machine learning and data analytics, and their applications to industrial process.
Name : Youngdoo Son
Affiliation : Dongguk University-Seoul,Korea
Biography : Youngdoo Son received the M.S. degree in industrial and management engineering from Pohang University of Science and Technology, Pohang, South Korea, in 2012, and the Ph.D. degree in industrial engineering from Seoul National University, Seoul, South Korea, in 2015. He is currently an Assistant Professor with the Department of Industrial and Systems Engineering, Dongguk University at Seoul, Seoul. His research interests include machine learning, neural networks, Bayesian methods, and their industrial and business applications.
Name : Jana Shafi
Affiliation : Prince Sattam bin Abdul Aziz University, KSA
Biography : Jana Shafi is a Lecturer at Prince Sattam bin Abdul Aziz University, KSA. She has published research papers in various International conferences and Journals. Her Research interests include Online Social Networks with technologies of Machine Learning, Deep Learning. She is a member of Elsevier Advisory Panel also a Mendeley Advisor.
Barkha Kakkar1, Prashant Johri1, Yogesh Kumar2, Hyunwoo Park3, Youngdoo Son3, and Jana Shafi4,*, An IoMT-Based Federated and Deep Transfer Learning Approach to the Detection of Diverse Chest Diseases Using Chest X-Rays, Article number: 12:24 (2022) Cite this article 1 Accesses
Download citationAnyone you share the following link with will be able to read this content:
Provided by the Springer Nature SharedIt content-sharing initiative