ArticlePDF Available

A Web-Based Platform for the Automatic Stratification of ARDS Severity

Authors:

Abstract and Figures

Acute respiratory distress syndrome (ARDS), including severe pulmonary COVID infection, is associated with a high mortality rate. It is crucial to detect ARDS early, as a late diagnosis may lead to serious complications in treatment. One of the challenges in ARDS diagnosis is chest X-ray (CXR) interpretation. ARDS causes diffuse infiltrates through the lungs that must be identified using chest radiography. In this paper, we present a web-based platform leveraging artificial intelligence (AI) to automatically assess pediatric ARDS (PARDS) using CXR images. Our system computes a severity score to identify and grade ARDS in CXR images. Moreover, the platform provides an image highlighting the lung fields, which can be utilized for prospective AI-based systems. A deep learning (DL) approach is employed to analyze the input data. A novel DL model, named Dense-Ynet, is trained using a CXR dataset in which clinical specialists previously labelled the two halves (upper and lower) of each lung. The assessment results show that our platform achieves a recall rate of 95.25% and a precision of 88.02%. The web platform, named PARDS-CxR, assigns severity scores to input CXR images that are compatible with current definitions of ARDS and PARDS. Once it has undergone external validation, PARDS-CxR will serve as an essential component in a clinical AI framework for diagnosing ARDS.
Content may be subject to copyright.
Citation: Yahyatabar, M.; Jouvet, P.;
Fily, D.; Rambaud, J; Levy, M;
Khemani, R.G.; Cheriet, F.,
on behalf of the Pediatric Acute
Respiratory Distress Syndrome
Incidence and Epidemiology
(PARDIE) V3 Investigators and
PALISI Network. A Web-Based
Platform for the Automatic
Stratification of ARDS Severity.
Diagnostics 2023,13, 933.
https://doi.org/10.3390/
diagnostics13050933
Academic Editors: Chiara Romei
and Emanuele Neri
Received: 15 January 2023
Revised: 23 February 2023
Accepted: 24 February 2023
Published: 1 March 2023
Copyright: © 2023 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
diagnostics
Article
A Web-Based Platform for the Automatic Stratification
of ARDS Severity
Mohammad Yahyatabar 1, Philippe Jouvet 2,, Donatien Fily 2, Jérome Rambaud 2, Michaël Levy 2,
Robinder G. Khemani 3, Farida Cheriet 1,† on behalf of the Pediatric Acute Respiratory Distress Syndrome
Incidence and Epidemiology (PARDIE) V3 Investigators and PALISI Network
1
Department of Computer and Software Engineering, Polytechnique Montréal, Montréal, QC H3T 1J4, Canada;
2Department of Pediatrics, Faculty of Medicine, University of Montréal, Montréal, QC H3C 3J7, Canada
3Department of Anesthesiology and Critical Care Medicine, Children’s Hospital of Los Angeles,
Los Angeles, CA 90027, USA
*Correspondence: philippe.jouvet@umontreal.ca
Membership of the Pediatric Acute Respiratory Distress Syndrome Incidence and Epidemiology (PARDIE) V3
Investigators and PALISI Network is provided in the Acknowledgments.
Abstract:
Acute respiratory distress syndrome (ARDS), including severe pulmonary COVID infection,
is associated with a high mortality rate. It is crucial to detect ARDS early, as a late diagnosis may
lead to serious complications in treatment. One of the challenges in ARDS diagnosis is chest X-ray
(CXR) interpretation. ARDS causes diffuse infiltrates through the lungs that must be identified using
chest radiography. In this paper, we present a web-based platform leveraging artificial intelligence
(AI) to automatically assess pediatric ARDS (PARDS) using CXR images. Our system computes a
severity score to identify and grade ARDS in CXR images. Moreover, the platform provides an image
highlighting the lung fields, which can be utilized for prospective AI-based systems. A deep learning
(DL) approach is employed to analyze the input data. A novel DL model, named Dense-Ynet, is
trained using a CXR dataset in which clinical specialists previously labelled the two halves (upper
and lower) of each lung. The assessment results show that our platform achieves a recall rate of
95.25% and a precision of 88.02%. The web platform, named PARDS-CxR, assigns severity scores
to input CXR images that are compatible with current definitions of ARDS and PARDS. Once it has
undergone external validation, PARDS-CxR will serve as an essential component in a clinical AI
framework for diagnosing ARDS.
Keywords:
chest X-ray; machine learning; acute respiratory distress syndrome; pediatric acute
respiratory distress syndrome; web-based platform
1. Introduction
Acute respiratory distress syndrome (ARDS) is a severe, even life-threatening condi-
tion, associated with respiratory failure, i.e., the inability of the lungs to fulfill their basic
function of exchanging gases in the body. ARDS occurs in children and adults; its main
causes include respiratory infection, aspiration, or trauma. The first description of ARDS as
a separate disease was provided in 1967. Variability in the ability to identify ARDS causes
difficulty in clinical trials. The Berlin definition introduced diagnostic criteria, such as acute
onset, severe hypoxemia (lack of oxygen in the blood), bilateral diffuse infiltrates visible
in chest radiography, and absence of any evidence of cardiac failure or fluid overload [
1
].
Despite intensive studies investigating ARDS (60,000+ articles found in PubMed), its mor-
tality rate is still as high as 43% [
2
]. Among the survivors of ARDS, a significant portion
experienced lasting damage to the lungs, especially in older patients. The Berlin definition
grades the severity of ARDS as being mild, moderate, or severe. Table 1illustrates the
oxygenation criteria and mortality rates associated with these severity levels.
Diagnostics 2023,13, 933. https://doi.org/10.3390/diagnostics13050933 https://www.mdpi.com/journal/diagnostics
Diagnostics 2023,13, 933 2 of 16
Table 1.
ARDS severities in the Berlin definition and associated oxygenation levels and mortality
rates [1].
Severity PaO2/FiO2Mortality
Mild 200–300 27%
Moderate 100–200 32%
Severe 100 45%
As seen in Table 1, considering the high mortality rate of ARDS and its rapid progres-
sion, early diagnosis of ARDS is vital. Furthermore, the mortality rate is directly associated
with the severity of the syndrome. The risk benefit profile of therapies depends on ARDS
severity, making early stratification of ARDS severity crucial for management. The Pediatric
Acute Lung Injury Consensus Conferences (PALICC) [
3
5
] were organized to address pedi-
atric ARDS (PARDS) specifications and give treatment and diagnosis recommendations.
According to the most recent definition of PARDS, PALICC-2 [
5
], the criteria allow for
new infiltrates in chest radiography, even if only a region within a single lung is affected.
One of the main reasons for this change in diagnostic criteria was the lack of agreement
in the interpretation of chest images between radiologists or between radiologists and
intensive care practitioners on the presence of bilateral infiltrates, which are required in the
Berlin standard. López-Fernández et al. showed that interobserver agreement for bilateral
infiltrates and quadrants of consolidation in PARDS was “slight”
(kappa 0.31 and 0.33) [6]
.
Sjoding et al. reported similar results, with interobserver reliability of ARDS diagnosis
being “moderate” (kappa = 0.50; 95CI, 0.40–0.59). Hence, there is an urgent need to improve
the reliability of Chest X-ray (CXR) intepretation in ARDS and PARDS to allow earlier
diagnosis of the syndrome [7].
Several studies have applied machine learning (ML) and artificial intelligence (AI)
approaches to analyze CXR images. One of the most common tasks reported in the literature
is diagnosing pulmonary pathologies using chest radiography. Thanks to massive publicly
available datasets, deep learning (DL) approaches have been broadly applied in medical
pathology detection. However, there is as yet no dataset annotated with ARDS labels. Thus,
few studies are found in the literature addressing the diagnosis of the syndrome.
To our knowledge, two papers present ML-based systems to identify ARDS in CXR
images. The first one [
8
] proposed a method for detecting ARDS using a traditional ML
approach based on hand-crafted features. The texture of intercostal image regions is
considered as a discriminative feature for classifying samples. To highlight intercostal areas,
a semi-automatic approach proposed by Plourde is utilized [
9
]. They succeed in reducing
the inter-observer variability between clinicians in diagnosing PARDS. However, their
approach is not automatic, and the rib segmentation step requires operator intervention.
In the second work, an automatic ARDS detection and grading approach was proposed
using a state-of-the-art DL model (Densenet) [
10
]. The authors first pretrained the model on
public datasets (not containing ARDS samples) and then refined the model with a custom
dataset consisting of ARDS-labeled images. Their approach performs well in diagnosing
ARDS, but the model provides no evidence for the support system’s decisions. Thus,
although it works well in analyzing ARDS cases, the model lacks interpretability, which is
essential for a ML system to be used in clinical settings.
Recently, due to the COVID-19 outbreak, the research community has gotten more
involved in computer-based analysis of chest X-ray images as one of the easiest and fastest
ways to check for signs of the disease. Mobile Chest X-ray Analysis [
11
] and Chester [
12
]
are prototype systems for CXR assessment developed using the aforementioned Densenet
model, trained on the public Chest-Xray14 dataset [
13
]. Both systems provide evidence
for the detected pathologies by means of saliency maps obtained using GradCam [
14
].
However, this can reveal areas that are irrelevant to the pathology being detected [
15
,
16
].
Thus, although these systems provide activation maps pointing out the references for the
decisions, they are not sufficiently reliable to be used in clinics.
Diagnostics 2023,13, 933 3 of 16
The main contributions of this paper are to create a tool for stratifying the severity
of ARDS in CXR images and to build a web-based platform for external validation. The
platform uses local information to classify X-rays based on the distribution of infiltrates
in the different lung quadrants, and it provides a global severity score for the image that
is applicable in both children and adults. The web-based platform, PARDS-CxR, can be
used as a standalone tool, or it can be integrated with other ARDS analysis tools to offer a
comprehensive approach for clinical use.
The following section first explains the details of the data collection used to train our
DL model. Then, we describe the proposed DL model and its evaluation process, and we
present the development of the Web platform. Section 3presents the results of testing the
ARDS assessment tool, and in Section 4, the strengths and drawbacks of our platform are
discussed. Section 5provides concluding remarks for this paper.
2. Materials and Methods
2.1. Methodology
This study contains four main phases, as illustrated in Figure 1. The end product is
PARDS-CxR, the web-based application to detect ARDS. First, a substantial set of data is
required to train the model. Existing public datasets do not include ARDS-labeled CXR
images, so we created a new one. This data collection process is summarized in Section 2.2.
Then, the proposed model must be trained on the CXR images. The model has two outputs
associated with lung segmentation and ARDS classification, as explained in Section 2.3.
The trained model is then tested on unseen data to be evaluated.
Sections 2.42.6
detail the
validation process. Finally, the model is uploaded to a server, and an interface is designed
so the user can easily access it. The web application is addressed in Section 2.7.
2.2. ARDS Dataset
Collectively, the main publicly available CXR datasets provide around a million images
with pathology labels [
17
]. This data motivated many researchers to employ AI techniques
in this domain. However, no such datasets assign ARDS-specific labels to images. As our
first step, we collected and annotated a dataset at Sainte-Justine Hospital, Montreal, Canada
(CHUSJ), to address the lack of appropriate data. Our dataset comprises three data sources
containing 373 CXR images. Ninety images and their corresponding labels came from a
previous study by our team [
8
]. A further 100 images were taken from the Chest X-ray14
dataset [
13
] and relabeled by clinical experts (JR, ML) in the hospital. Another 183 images
were provided by the PARDIE study, a multi-national study that prospectively gathered
chest X-ray images of children with ARDS [
6
]. For each image, labels were associated with
the four lung quadrants obtained by splitting each lung into upper and lower portions. We
refer to each quadrant by its position: left upper (LR), left lower (LL), right upper (RU),
and right lower (RL). According to the Berlin definition [
1
], visible bilateral infiltrates are a
mandatory criterion for a case to be categorized as ARDS. Two intensivists from CHUSJ
assessed the presence of infiltrates in each quadrant. A sample was included in the dataset
only if the clinical observers reached a consensus on the labels. In addition, 138 CXR images
were taken from the Montgomery dataset to represent the normal class [
18
]. These samples
were labeled as non-ARDS if agreed by the clinical experts. By dropping images with
disagreements in their labeling, our final ARDS dataset consisted of 356 images, of which
134 meet the bilateral infiltrates criteria in the Berlin definition [1].
Diagnostics 2023,13, 933 4 of 16
RL LL
RU
Model Training
L RU
L LU
L RL
L LL
ARDS DATASET
Joint Segmentation-Classification
Dense Y-net
L RU
L LU
L RL
L LL
+
LossClassification
+
LossSegmentation
Loss =
Trained Dense Y-net
1
0
1
1
Bi-lateral infiltrates? ARDS (YES / NO)
Training Data
Testing Data
Target
Data collection
Evaluation
Trained Dense Y-net UI Presentation
Back-End
Web Application
Server Side User Side
Front-End
Clinical Review
PARDS-CxR
LU
Figure 1. Organization of our study into four main phases. The data are collected from several data
sources and annotated at Saint-Justine Hospital, Montreal, Canada. The DL model is trained using
quadrant-level labels and lung segmentation maps. It is then evaluated on a set of previously unseen
images; both the classification and segmentation performances are assessed. Finally, a web-based
platform is designed and made available through the internet.
Diagnostics 2023,13, 933 5 of 16
2.3. Joint Segmentation and Classification Model
In computer-based diagnosis approaches, it is common to use segmentation ahead of
classification to determine the region of interest. Lung segmentation separates the lung
areas from the thoracic tissues surrounding them and is the primary image analysis step
in many clinical decision support systems. Generalization to new datasets is a difficult
challenge in the analysis of chest radiography. In that respect, segmentation is considered a
strategy to limit the impact of specific imaging devices and settings, since it restricts the
feature extraction to the lung fields and removes the effect of the image background [
19
,
20
].
However, serial usage of segmentation and classification propagates the segmentation
error into the classification network. Dense-Ynet is a convolutional network that takes
advantage of Densenet, Y-net, and U-net models to do both tasks simultaneously in a joint
segmentation–classification model. The backbone of the network used in this study is our
previously developed Dense-Unet [
21
]. Dense-Unet is a segmentation model in which
dense connections between the feature maps in various layers facilitate the information
flow throughout the model, letting designers choose a configuration with a small number
of training parameters. Our proposed Dense-Ynet takes advantage of automatic feature
extraction from both the original and segmented images (Figure 2). The model has two
outputs and is trained using two loss functions: the lung segmentation loss and quadrant
classification loss. The model works based on the convolution operation. A convolution
is a mathematical operation that filters the information of its input and creates feature
maps. An inevitable effect of the convolution operation is to change the dimensions of the
feature maps. To tackle this issue, upsampling and strided convolution operations are used
to ensure that feature maps coming from different layers can be concatenated. Squeeze
and excitation (SE) blocks [
22
] are also used after each convolution layer to improve the
representational power of the blocks by recalibrating the features. The key strengths of
Dense-Ynet are use of lung segmentation in its architecture, specialized connectivity, which
enable better generalization, and prediction of local labels for each image.
To reach the final decision based on the Berlin definition, we must test for existing
bilateral infiltrates. To that end, a simple logical operation in Equation (1) combines the
predictions of each quadrant to check this condition:
PARDS = (PRL PRU )(PLL PLU)(1)
PRL
,
PRU
,
PLL
, and
PLU
are the prediction labels for the right lower, right upper, left
lower, and left upper quadrants, respectively.
PARDS
is the inferred ARDS label, and
and
are logical or and and operations. The equation states that, if at least one quadrant is
involved on each side, the case is recognized as (P)ARDS.
Dense-Ynet
Network LABELRU
LABELLU
LABELLL
LABELRL
ARDS(Y/N)
Input
Deep model Lung Segmentation
ARDS Classification
Figure 2.
The Dense-Ynet model takes advantage of the interaction between the segmentation
and classification tasks by performing them simultaneously. The features from the original and
lung-segmented images are concatenated and utilized to classify ARDS cases.
Diagnostics 2023,13, 933 6 of 16
2.4. Experimental Design
In this work, 267 images of the ARDS dataset are used to train the Dense-Ynet model.
In addition, 35 images are used to validate the training process. For the testing stage,
54 images previously unseen by the network are used. The algorithm is evaluated with the
five-fold cross-validation strategy. Cross-validation is a method that tries various training
and testing data combinations to confirm the reported results’ reliability. Data augmentation
is a technique to enrich the training data by generating new images from the current training
set. For this purpose, we use basic image processing techniques, such as random rotation,
cropping, shifting, horizontal flipping, and intensity changing. The rectified linear unit
(
ReLu
) activation function introduces non-linearity to network blocks. The
Sigmoid
function
provides valid labels between zero and one in both the segmentation and classification
output layers.
Adam
is the optimizer used for updating the model weights during training.
To reach the optimal configuration, a set of hyperparameters must be explored to find the
best model structure and training policy. The Web platform
(see Section 2.7)
employs six
Dense-Ynet instances, corresponding to the best hyper-parameters sets. Using an ensemble
approach, the final result presented to the user combines the values received from the
individual models.
The PARDS-CxR application detects lung quadrants consolidation, and the final ARDS
label is derived from the quadrant predictions using Equation (1).
2.5. Scoring Scheme
To analyze the severity of ARDS in CXR images, a scoring scheme is proposed based
on the number and the position of affected lung quadrants (see Table 2). The scheme is
compatible with the Berlin definition, in which existing bilateral infiltrates are an essential
criterion for ARDS diagnosis in chest radiography.
Table 2. Severity scoring scheme based on affected lung quadrants.
Affected Quadrants Score Severity
4 quadrants 5 Severe
3 quadrants 4
2 quadrants (Different sides) 3 Mild
2 quadrants (Same side) 2
Non-ARDS
1 quadrant 1
No affected quadrant 0
Giving scores is important from two points of view. First, the score represents the
severity of the diffused infiltrates throughout the lungs. Second, reporting disease severity
helps clinicians follow appropriate treatment protocols or triaging. This type of system has
been proposed for the Murray Lung Injury Score, as well as as part of the recently proposed
RALE score in adult patients with ARDS.
2.6. Evaluation Metrics
Evaluation metrics are measured from the algorithm’s performance on unseen test data
to assess the approach. There is no metric representing the total capacity of the PARDS-CxR
platform. However, we use a set of performance metrics to provide a complete overview of
the model’s operation. A confusion matrix quantifies the ability of the classifier to detect
each class separately. It gives detailed measures comparing the actual and predicted labels,
as shown in Figure 3.
Diagnostics 2023,13, 933 7 of 16
Figure 3.
Confusion matrix for a binary classification problem. The matrix contains four elements
that, together, evaluate the system’s predictions versus the real labels.
The elements of the confusion matrix, namely, the true positive (TP), true negative (TN),
false positive (FP), and false negative (FN) values, serve to calculate several assessment
metrics as follows:
Accuracy =TP +TN
FP +FN +TP +TN (2)
Precision =TP
FP +TP (3)
Recall =T P
FN +T P (4)
F1=2×Precision ×Recall
Precision +Recall . (5)
The
Accuracy
metric represents the overall correctness of a classification algorithm. It
cannot fully express the model performance, however, especially in the case of unbalanced
testing data.
Precision
and
Recall
reveal the model’s performance in discriminating between
the different classes.
Precision
represents how precise the model is in identifying the target
(positive) class. Specifically, it points out what portion of cases predicted as positive are
really ARDS cases. On the other hand, the
Recall
value shows what proportion of predicted
ARDS cases are actually labeled as ARDS. These two metrics have a complementary role in
describing the model’s behavior. The
F
1 score, derived from
Precision
and
Recall
values, is
a single metric to quantify the algorithm’s performance.
The receiver operating characteristic (ROC) curve illustrates the diagnostic capacity of
a system by comparing true positive and false positive rates as the discrimination threshold
(applied at the network’s output layer to decide between the two classes) varies. The area
under the ROC curve (AUROC) represents the discriminatory power of the classifier.
2.7. Web-Based Platform
We designed a web-based platform to facilitate the diagnosis of ARDS in CXR images
by medical professionals. The platform is intended as a tool to provide a second opinion
to clinicians, but no direct medical use is recommended until medical professionals validate
the tool using external data. The PARDS-CxR platform takes advantage of six Dense-Ynet
instances to provide scores for each input image. The scores are given based on the number and
combination of affected lung quadrants as explained in Section 2.5. A global score is assigned by
combining the outputs from the model instances. In addition, the application provides accurate
lung segmentation maps, which are helpful in AI-based analysis of CXR images.
The web application utilizes the React library to create a user-friendly and interactive
user interface (UI) for delivering the specified services. The library enables efficient code
writing and makes it easier to manage, refine, and integrate the application with other tools.
The platform supports both English and French languages and has two main modes for
ARDS definitions for adults (Berlin) and children (PALICC-2). The difference between the
modes is that, when using PALICC-2 mode, the platform requires two input images. The
application response includes segmentation maps, severity scores (local and global), and
an interpretation based on the definition.
Diagnostics 2023,13, 933 8 of 16
Although the deep models are trained using graphical processing units (GPUs), the
evaluation model does not require a GPU and can process the results in 2-3 s. Thus, the
running bottleneck could be the network connection speed. The application is capable
of storing data and providing log files, but this feature is currently disabled and will be
activated when the validation protocol is approved. The PARDS-CxR platform is detailed
further in Section 3.3.
3. Results
3.1. Quadrant-Based Classification
The PARDS-CxR web-based platform uses Dense-Ynet as the joint segmentation-
classification model. In classification, the model predicts four labels associated with lung
quadrants, as explained in Section 2.3. The platform uses an ensemble of six Dense-Ynet
model instances with different training and model structure configurations. Regarding
model structures, we experimented with different channel depths in convolution blocks,
loss functions, weights for merging loss functions, activation functions, and initial network
weights. For the training configurations, we varied several hyperparameters, namely, the
learning rate, training batch size, augmentation probability, and stopping criterion.
Figure 4shows the confusion matrix of the ensemble of models. To merge the results
from the model instances, a hard voting strategy is employed based on the labels predicted
independently by the models. To be precise, each model is trained separately with its
specific configuration. The testing is also done independently, and if at least three models
decide that an image is an ARDS case, the combined result is positive. By combining
models with various configurations, the intrinsic biases of each one to accept or reject an
image as ARDS are balanced in the ensemble output. Thus, the final performance improves
compared to any individual model.
Non-ARDS ARDS
Actual
Non-ARDS
ARDS
Predicted
64.1% 1.5%
4.1% 30.3%
Figure 4.
Final confusion matrix obtained from the combination of network instances using hard
voting. The numbers (percentages) are obtained by taking the average of several tests (five-fold
cross validation).
Table 3compares the classification performances of the Dense-Ynet instances in terms of
the four assessment metrics seen previously. Some of the listed models achieve higher precision,
while others reach better recall values. By combining the predicted labels provided by these
models, the ensemble algorithm achieves the highest
F
1 score, representing the best compromise
between precision and recall. Indeed, ensembling the models does not outperform every one in
terms of Precision and Recall, but the final F1 and accuracy values improve.
Diagnostics 2023,13, 933 9 of 16
Table 3.
Evaluation of the six models and the result of their combination (ensemble model) for classification.
Model Accuracy Recall Precision F1
Network 1 92.95% 88.45% 91.99% 90.19%
Network 2 93.54% 96.41% 84.37% 89.99%
Network 3 92.04% 94.42% 87.89% 91.03%
Network 4 92.96% 100.0% 83.33% 90.91%
Network 5 87.32% 100.0% 74.29% 85.25%
Network 6 88.74% 80.01% 80.01% 80.02%
Ensemble model 94.35% 95.25% 88.02% 91.49%
In this paper, the problem of ARDS diagnosis is based on the classification of lung
quadrants. Thus, the task can also be considered as a multi-label classification problem.
Figure 5
shows the ROC curves of all quadrants’ predictions for the Dense-Ynet instances,
i.e., the ROC curves associated with the binary classification of the lung quadrants, re-
gardless of their positions. The AUROC metric is not directly related to the system’s
performance in ARDS diagnosis, but the misclassification of one lung quadrant may cause
an error in classifying the image as a whole.
Figure 5. ROC curves for classification of lung quadrants regardless of their position in the lungs.
3.2. ARDS Severity Prediction
As seen in Table 2, the application determines the severity of ARDS in CXR images
based on the number and combination of affected lung quadrants. The platform provides
a global score for each input image by taking the average of the scores from each model.
CXR images are then categorized into one of three severity grades based on the predicted
scores: non-ARDS, mild ARDS, and severe ARDS. The platform’s effectiveness in deter-
mining ARDS severity is illustrated in Figure 6. The three-class confusion matrix shows
that the approach can detect ARDS and discriminate between mild and severe states of
the syndrome.
Diagnostics 2023,13, 933 10 of 16
Non-ARDS Mild ARDS Severe ARDS
Actual
Non-ARDS
Mild ARDS
Severe ARDS
Predicted
64.1% 0.5% 0.2%
1.5% 11.5% 0.7%
1.8% 0.9% 18.8%
Figure 6.
Confusion matrix for classification of ARDS severity with three levels (none, mild, severe).
3.3. PARDS-CxR, the Web-Based Platform
Our web application is currently loaded on a web server at CHUSJ and is accessible
at the address (https://chestxray-reader.chusj-sip-ia.ca, accessed on 15 January 2023).
The process of training and testing the deep model was programmed in Python using the
PyTorch library [
23
]. The training process and hyperparameter search were executed on GPU,
as they required intensive parallel computing. The trained model was then transferred to CPU
to evaluate new images; thus, no graphical processor is necessary on the server to run the
application. The graphical user interface was written in JavaScript and is compatible with
various internet browsers on the client side. No data are kept on the server side, and the
application output image is available to store in the user’s local storage. The user interface
works in English and French, and CXR images can be uploaded using the menu option or
drag-and-drop (see Figure 7).
The application bases itself on the most accepted definitions for ARDS and PARDS.
Based on the Berlin definition, the presence of bilateral infiltrates in chest radiography is
a criterion manifesting the existence of ARDS [
1
]. The platform processes the image and
displays its decision by providing a percentage associated with the level of infiltration
in each quadrant (Figure 7). A global percentage is also given based on infiltrate levels
of infiltrates in quadrants and their combination as in Table 2. This value represents the
severity of ARDS in the input image. An image with a global percentage above 60%
is interpreted as an ARDS case, since, based on the proposed severity scoring system,
infiltrates should be diffused through both lungs. Reporting each quadrant’s involvement
is necessary, since it gives the rationale behind the global severity measure. As seen in
Figure 7, a segmentation map highlighting the lung segments is also provided.
Identifying progression of ARDS is also possible, as two images taken at different times can
be compared by the system. Additionally, an example of CXR image comparison is displayed
in Figure 8.
Diagnostics 2023,13, 933 11 of 16
PARDS-CxR
Dashboard
Profile
chusj-research
Mode
ARDS
PARDS
Documentation
Help
References
About
Quadrant UR
83.3 %
Quadrant UL
33.3 %
Quadrant LR
83.3 %
Quadrant LL
16.7 %
Lung Mask Regional Severity
The CXR image does not meet ARDS Criterion (Berlin definition)
50 %
Download Result Upload New Image Upload second Image
(Switch Mode)
Figure 7.
Main interface of the PARDS-CxR web application. In the standard mode, a single CXR
image is analyzed according to the Berlin definition.
PARDS-CxR
Dashboard
Profile
chusj-research
Mode
ARDS
PARDS
Documentation
Help
References
About Lung Mask
Progressive infiltrates detected (50% to 100%)
CXR images meet PARDS Criterion (PALICC-2 definition)
Download Result Upload New Image
Regional Severity
Quadrant UR
83.3 %
Quadrant UL
33.3 %
Quadrant LR
83.3 %
Quadrant LL
16.7 %
Quadrant UR
100 %
Quadrant UL
100 %
Quadrant LR
100 %
Quadrant LL
100 %
Figure 8.
PARDS-CxR interface in image comparison mode. The platform can analyze two CXR
images to detect ARDS progression based on the PALICC-2 definition.
Diagnostics 2023,13, 933 12 of 16
4. Discussion
The proposed DenseY-net is a joint segmentation–classification model that diagnoses
(P)ARDS based on lung quadrant-level classification. The results show that the model can
accurately classify quadrants and, consequently, the entire input image. This labeling strat-
egy offers a reasoning framework for decision-making and incorporates an interpretability
feature into the platform. Ensemble modeling is used to combine the outcomes from six
model instances. PARDS-CxR can also do lung field segmentation, which is a necessary
element in many decision support systems. Our approach performs well in detecting the
severity of ARDS by giving a score to each input determined by the number and posi-
tion of affected lung quadrants. This makes the model compatible with both ARDS and
PARDS definitions.
A few large chest radiography datasets are publicly available for the research commu-
nity [
13
,
24
,
25
]. A key benefit of deep learning is its capacity to analyze and learn features
from a substantial amount of data. Therefore, it is unsurprising that several ML researchers
have investigated CXR image analysis in various contexts. However, important limitations
of these datasets make them unsuitable for developing dependable systems for the hospital
setting. Indeed, most of the data are annotated using clinicians’ notes processed by natural
language processing (NLP) techniques [
26
]. This leads to erroneous labeling of a portion of
the images. For example, a 10% error is reported for Chest X-ray 14 [13], even though it is
one of the most frequently used CXR datasets. The clinical review in [
27
] reveals an even
higher rate of data labeling errors in that dataset.
Although adding some level of noise to the training inputs can improve a deep model’s
performance, biases and extensive labeling errors will decrease the model’s accuracy. This
could be a reason for the relatively poor generalization ability of deep models when
confronting new samples from other data sources. Furthermore, available samples are
annotated for a limited number of pathologies. Public CXR datasets cover between 14 and
18 chest pathologies, but these do not include ARDS or PARDS. To address this constraint,
we collected our own CXR dataset from three different sources and annotated it for PARDS
at CHUSJ. This dataset was labeled at the lung quadrant level, and the lung fields were
manually identified in each image to establish a segmentation ground truth. The resulting
dataset contains 356 CXR images, including 134 that meet the bilateral criteria for ARDS.
Annotating data is costly in the clinical field, even more so considering that the DenseY-
net model needs lung maps and quadrant-level ground-truth labels. Consequently, our
ARDS dataset is relatively small. Nonetheless, our model is designed in such a way as to
train adequately on small datasets. The specialized connectivity within the model allows
for the creation of a lighter model with shallower intermediate feature maps, resulting
in a smaller number of training parameters. A model with fewer parameters is more
appropriate for training with small datasets. The algorithm was assessed on our own
dataset, as explained in Section 2.4. A bigger dataset could increase the generalization
capacity of the model ensemble. Moreover, external validation of the platform using data
from various health centers will make it more reliable as a tool for prospective clinical
research. Thus, as next steps in the web application’s development, external validation
and improving interpretability are two major points, since both are necessary to turn the
platform into a practical tool in clinics.
Moreover, according to the (P)ARDS definition, co-occurrence of detectable infiltrates
in CXR and hypoxemia is necessary when no evidence of cardiogenic pulmonary edema
is observed. Thus, although the presence of infiltrates in chest radiography is known as
the most limiting factor for diagnosing ARDS, meeting other criteria is a requisite. The
Clinical Decision Support System (CDSS) lab at CHUSJ has the capacity to investigate other
ARDS diagnosis criteria, including cardiac failure and hypoxemia. Le et al. have employed
NLP techniques and ML algorithms to detect cardiac failure in children [
28
]. Sauthier et al.
have developed a method to accurately estimate Pao2 levels using noninvasive data [
29
].
Integrating the tool proposed in this study with other works will lead to a system giving
comprehensive ARDS diagnoses. Sufficient electronic medical infrastructure is available in
Diagnostics 2023,13, 933 13 of 16
the PICU of CHUSJ to facilitate the flow of data from various sources [
30
]. By accessing data
from clinical narrative analysis, measuring oxygenation indices, and detecting infiltrates
in CXR images, it will be possible to make clinical decisions in real time. Therefore, an
important objective for our team will be to implement an ARDS diagnosis package at
CHUSJ, integrating all these criteria and data sources.
The strength of this study lies in the development of an algorithm that, in comparison
to existing approaches, is more interpretable and automated and is compatible with existing
ARDS definitions. Unlike an earlier ARDS diagnosis method proposed by our research
team [
8
], the DL-based approach used in this application does not need any interaction from
clinicians or operators to guide the algorithm. The novel model provides an end-to-end
process that is simple for the user and provides the diagnotic outputs instantaneously.
Recently, Sjoding et al. [
10
] proposed annother automatic algorithm for detecting ARDS in
CXR images. However, their approach lacks explainability, i.e., the system’s decisions are
not supported by further information. By contract, since PARDS-CxR detects infiltrates in
each lung quadrant, the basis for the decision is integral to our method. This strengthens
the platform’s reliability, since the user can reject or accept the decision by observing the
delivered explanation. In addition, the proposed approach is compatible with both PARDS
and ARDS definitions [
1
,
3
], as the scoring scheme used translates to a disease severity level.
At present, the main limitation of our algorithm is its lack of external validation. Indeed,
its development relied on a limited number of CXR images with a single team annotating
them. For this reason, we have implemented the algorithm on a web platform to allow
researchers to conduct validation studies.
5. Conclusions
This work has described a deep learning method and web-based platform for diag-
nosing acute respiratory distress syndrome (ARDS) from chest X-ray (CXR) images. The
platform uses an ensemble of novel Dense-Ynet networks that can accurately detect lung
infiltrates in different quadrants and combine this information to detect ARDS and grade
its severity. This approach ensures that our tool is compatible with various ARDS defini-
tions in both adults and children. Following feedback from clinical researchers during a
validation phase, the platform will be integrated into a complete clinical decision system
for ARDS. The tool presented here will serve as the CXR analysis component within an
AI-based framework that will monitor other factors, such as hypoxemia and occurence of
cardiac arrest.
Author Contributions:
P.J., F.C., M.Y. and R.G.K. conceptualized and designed the study. M.Y., P.J., F.C.,
D.F., M.L. and J.R. developed the study protocol. M.Y., D.F., M.L. and J.R. conducted the algorithm
development. M.Y., F.C. and P.J. drafted the initial manuscript. All authors approved the final manuscript
as submitted. All authors have read and agreed to the published version of the manuscript.
Funding:
This study was supported by grants from IVADO (Artificial Intelligence Research and
Transfer Institute), the Quebec Ministry of Health and Sainte-Justine Hospital. M.Y. is financed by
an award from the Fonds de Recherche en Santé du Québec (FRQS) Chair in Artificial Intelligence and
Health. P.J. earns a research salary from FRQS.
Institutional Review Board Statement:
The study was approved by the Institutional Review Board
of Sainte-Justine Hospital (approval number: 2023-5124).
Informed Consent Statement:
The study was carried out on a research database and the Institutional
Review Board did not require informed consent.
Data Availability Statement:
Access to data can be requested from Philippe Jouvet. Specific institu-
tional review board rules will apply.
Acknowledgments:
The authors gratefully acknowledge Philippe Debanné for his assistance in
reviewing the manuscript. Furthermore, the authors thank all the investigators (pediatric inten-
sivists/Radiologists) who participated in the PARDIE V3 study (Country, site and investigator
list): Argentina. Hospital De Ninos Ricardo Gutierrez: Rossana Poterala; Hospital de Ninos sor
Diagnostics 2023,13, 933 14 of 16
Maria Ludovica: Pablo Castellani/Martin Giampieri/Claudia Pedraza; Hospital Nacional Alejan-
dro Posadas: Nilda Agueda Vidal/Deheza Rosemary/Gonzalo Turon/Cecilia Monjes; Hospital
Pediatrico Juan Pablo II: Segundo F. Espanol; Hospital Universitario Austral: Alejandro Siaba Ser-
rate/Thomas Iolster/Silvio Torres; Sanatorio de Ninos de Rosario: Fernando Paziencia. Australia.
Princess Margaret Hospital for Children: Simon Erickson/Samantha Barr/Sara Shea. Bolivia. Hospi-
tal del Nino Manuel Ascencio Villaroel: Alejandro F. Martinez Leon/Gustavo A. Guzman Rivera.
Canada. CHU Sainte-Justine: Philippe Jouvet/Guillaume Emeriaud/Mariana Dumitrascu/Mary
Ellen French. Chile. Hospital Base de Valdivia: Daniel Caro I/Andrés A Retamal Caro; Hospital
El Carmen de Maipu: Pablo Cruces Romero/Tania Medina; Hospital Luis Calvo Mackenna: Car-
los Acuna; Hospital Padre Hurtado: Franco Diaz/Maria Jose Nunez. China. Children’s Hospital
of Fudan Univ: Yang Chen. Colombia. Clinica Infantil de Colsubsidio: Rosalba Pardo Carrero;
Hospital General de Medellin: Yurika P. Lopez Alarcon; Hospital Militar Central: Ledys María
Izquierdo; Hospital Pablo Tobon Uribe (HPTU): Byron E. Piñeres Olave. France. CHU de Nantes:
Pierre Bourgoin; Hopital d’enfants de Brabois–CHU de Nancy: Matthieu Maria. Greece. University
of Crete, University Hospital PICU: George Briassoulis/Stavroula Ilia. Italy. Children’s Hospital
Bambino Gesu: Matteo Di Nardo/Fabrizio Chiusolo/Ilaria Erba/Orsola Gawronski; Children’s
Hospital Vittore Buzzi: Anna Camporesi. Japan. Hiroshima University: Nobuaki Shime/Shinichiro
Ohshimo/Yoshiko Kida/Michihito Kyo. Malaysia. Universiti Kebangsaan Malaysia: Swee Fong
Tang/Chian Wern Tai; University Malaya Medical Center: Lucy Chai See Lum/Ismail Elghuwael.
Mexico. Hospital Espanol De Mexico: Nestor J. Jimenez Rivera. Peru. Hospital de Emergencias
Pediatricas: Daniel Vasquez Miranda/Grimaldo Ramirez Cortez; Instituto Nacional de Salud del
Nino: Jose Tantalean. Portugal. Hospital Santa Maria–Centro Hospitalar Lisboa Norte: Cristina
Camilo. Saudi Arabia. King Abdullah Specialist Children’s Hospital, King Abdulaziz Medical City:
Tarek Hazwani/Nedaa Aldairi/Ahmed Al Amoudi/Ahmad Alahmadti. Spain. Cruces University
Hospital: Yolanda Lopez Fernandez/Juan Ramon Valle/Lidia Martinez/Javier Pilar Orive; Hospi-
tal Regional Universitario de Malaga: Jose Manuel Gonzalez Gomez/Antonio Morales Martinez;
Hospital Universitari I Politecnic La Fe: Vicent Modesto I Alapont; Sant Joan de Deu University
Hospital: Marti Pons Odena; Hospital Universitario Central De Asturias: Alberto Medina; Virgen
de la Arrixaca University Hospital: Susana Reyes Dominguez. Turkey. Akdeniz University School
of Medicine: Oguz Dursun/Ebru Atike Ongun; Izmir Katip Celebi University Medical School and
Tepecik Research and Training Hospital: Fulya Kamit Can/Ayse Berna Anil. UK. Evelina London
Children’s Hospital: Jon Lillie/Shane Tibby/Paul Wellman/Holly Belfield/Claire Lloyd; Great Or-
mond St. Children’s Hospital: Joe Brierley/Troy E. Dominguez/Eugenia Abaleke/Yael Feinstein;
Noah’s Ark Children’s Hospital for Wales: Siva Oruganti/Sara Harrison; Nottingham University
Hospitals: Catarina Silvestre; Oxford Radcliffe Hospitals NHS Foundation Trust: James Weitz; Royal
Manchester Children’s Hospital: Peter-Marc Fortune/Gayathri Subramanian/Claire Jennings; St.
Mary’s Hospital: David Inwald/Calandra Feather/May-Ai Seah/Joanna Danin. USA. Arkansas Chil-
dren’s Hospital: Ron Sanders/ Glenda Hefley/Katherine Irby/Lauren Edwards/Robert F Buchmann;
Children’s Hospital and Medical Center: Sidharth Mahapatra/Edward Truemper/Lucinda Kustka;
Children’s Hospital at Dartmouth: Sholeen T. Nett/Marcy Singleton/J. Dean Jarvis; Children’s
Hospital Colorado: Aline B. Maddux/Peter M. Mourani/Kimberly Ralston/Yamila Sierra/Jason
Weinman/Zach VanRheen/Christopher Newman; Children’s Hospital Los Angeles: Robinder Khe-
mani/Christopher Newth/Jeni Kwok/Rica Morzov/Natalie Mahieu; Children’s Hospital of Philadel-
phia: Nadir Yehya/Natalie Napolitano/Marie Murphy/Laurie Ronan/Ryan Morgan/Sherri Ku-
bis/Elizabeth Broden; Children’s Hospital of Wisconsin: Rainer Gedeit/Kathy Murkowski/Katherine
Woods/Mary Kasch; Children’s Mercy Hospital and Clinics: Yong Y. Han/Jeremy T. Affolter/Kelly S.
Tieves/Amber Hughes-Schalk; Cincinnati Children’s Hospital Medical Center: Ranjit S. Chima/Kelli
Krallman/Erin Stoneman/Laura Benken/Toni Yunger; Connecticut Children’s Medical Center:
Christopher L Carroll/James Santanelli; Inova Children’s Hospital: W. Keith Dockery/Shirin Jafari-
Namin/Dana Barry/Keary Jane’t; Joseph M Sanzari Children’s Hospital at Hackensack University
Medical Center: Shira Gertz; Nicklaus Children’s Hospital: Fernando Beltramo/Balagangadhar
Totapally/Beatriz Govantes; Northwestern University, Ann & Robert H Lurie Children’s Hospital
of Chicago: Bria Coates/Lawren Wellisch/Kiona Allen/Avani Shukla; Penn State Hershey Chil-
dren’s Hospital: Neal J. Thomas/Debbie Spear; Rainbow Babies and Children’s Hospital, Steven
L. Shein/Pauravi Vasavada; Saint Barnabas Medical Center: Shira Gertz; Stony Brook Children’s
Hospital: Margaret M. Parker/Daniel Sloniewsky; The Children’s Hospital of Oklahoma; Chris-
tine Allen/Amy Harrell; UCSF Benioff Children’s Hospital Oakland: Natalie Cvijanovich; Uni-
versity of Miami/Holtz Children’s Hospital: Asumthia S. Jeyapalan/Alvaro Coronado-Munoz;
Diagnostics 2023,13, 933 15 of 16
University of Michigan–C.S. Mott Children’s Hospital: Heidi Flori/Mary K. Dahmer/Chaandini Jay-
achandran/Joseph Kohne; University of Minnesota Masonic Children’s Hospital: Janet Hume/Dan
Nerheim/Kelly Dietz; University of WA/Seattle Children’s Hospital: Lincoln Smith/Silvia Hart-
mann/Erin Sullivan/Courtney Merritt; Weill Cornell Medical College: Deyin D. Hsing/Steve
Pon/Jim Brian Estil/Richa Gautam; Yale School of Medicine: John S. Giuliano Jr./Joana Ta.
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
AI Artificial Intelligence
ARDS Acute Respiratory distress Syndrome
AUROC Area Under the ROC Curve
CDSS Clinical Decision Support System
CHUSJ Centre Hospitalier Universitaire Sainte-Justine (Sainte-Justine Hospital)
CXR Chest X-ray
CPU Central Processing Unit
DL Deep Learning
GPU Graphical Processing Unit
LL Left Lower
LU Left Upper
ML Machine Learning
NLP Natural Language Processing
PALICC Pediatric Acute Lung Injury Consensus Conference
PARDS Pediatric Acute Respiratory Distress Syndrome
PICU Pediatric Intensive Care Unit
ReLU Rectified Linear Unit
RL Right Lower
ROC Receiver Operating Characteristic
RU Right Upper
UI User Interface
References
1.
Force, A.D.T.; Ranieri, V.; Rubenfeld, G.; Thompson, B.; Ferguson, N.; Caldwell, E.; Fan, E.; Camporota, L.; Slutsky, A. Acute
respiratory distress syndrome. JAMA 2012,307, 2526–2533.
2.
Sedhai, Y.R.; Yuan, M.; Ketcham, S.W.; Co, I.; Claar, D.D.; McSparron, J.I.; Prescott, H.C.; Sjoding, M.W. Validating measures of
disease severity in acute respiratory distress syndrome. Ann. Am. Thorac. Soc. 2021,18, 1211–1218. [CrossRef]
3.
Pediatric Acute Lung Injury Consensus Conference Group. Pediatric acute respiratory distress syndrome: Consensus recommen-
dations from the Pediatric Acute Lung Injury Consensus Conference. Pediatr. Crit. Care Med. J. Soc. Crit. Care Med. World Fed.
Pediatr. Intensive Crit. Care Soc. 2015,16, 428.
4.
Khemani, R.G.; Smith, L.; Lopez-Fernandez, Y.M.; Kwok, J.; Morzov, R.; Klein, M.J.; Yehya, N.; Willson, D.; Kneyber, M.C.; Lillie,
J.; et al. Paediatric acute respiratory distress syndrome incidence and epidemiology (PARDIE): An international, observational
study. Lancet Respir. Med. 2019,7, 115–128. [CrossRef]
5.
Emeriaud, G.; López-Fernández, Y.M.; Iyer, N.P.; Bembea, M.M.; Agulnik, A.; Barbaro, R.P.; Baudin, F.; Bhalla, A.; de Carvalho,
W.B.; Carroll, C.L.; et al. Executive Summary of the Second International Guidelines for the Diagnosis and Management of
Pediatric Acute Respiratory Distress Syndrome (PALICC-2). Pediatr. Crit. Care Med. 2023,24, 143–168. [CrossRef] [PubMed]
6.
López-Fernández, Y.M.; Smith, L.S.; Kohne, J.G.; Weinman, J.P.; Modesto-Alapont, V.; Reyes-Dominguez, S.B.; Medina, A.;
Piñeres-Olave, B.E.; Mahieu, N.; Klein, M.J.; et al. Prognostic relevance and inter-observer reliability of chest-imaging in pediatric
ARDS: A pediatric acute respiratory distress incidence and epidemiology (PARDIE) study. Intensive Care Med.
2020
,46, 1382–1393.
[CrossRef]
7.
Sjoding, M.W.; Hofer, T.P.; Co, I.; Courey, A.; Cooke, C.R.; Iwashyna, T.J. Interobserver reliability of the Berlin ARDS definition
and strategies to improve the reliability of ARDS diagnosis. Chest 2018,153, 361–367. [CrossRef] [PubMed]
8.
Zaglam, N.; Jouvet, P.; Flechelles, O.; Emeriaud, G.; Cheriet, F. Computer-aided diagnosis system for the Acute Respiratory
Distress Syndrome from chest radiographs. Comput. Biol. Med. 2014,52, 41–48. [CrossRef]
9.
Plourde, F.; Cheriet, F.; Dansereau, J. Semiautomatic Detection of Scoliotic Rib Borders From Posteroanterior Chest Radiographs.
IEEE Trans. Biomed. Eng. 2012,59, 909–919. [CrossRef] [PubMed]
Diagnostics 2023,13, 933 16 of 16
10.
Sjoding, M.W.; Taylor, D.; Motyka, J.; Lee, E.; Co, I.; Claar, D.; McSparron, J.I.; Ansari, S.; Kerlin, M.P.; Reilly, J.P.; et al. Deep
learning to detect acute respiratory distress syndrome on chest radiographs: A retrospective study with external validation.
Lancet Digit. Health 2021,3, e340–e348. [CrossRef] [PubMed]
11.
Kexugit. Using Microsoft AI to Build a Lung-Disease Prediction Model Using Chest X-ray Images. 2018. Available online:
https://learn.microsoft.com/en-ca/archive/blogs/machinelearning/using-microsoft-ai-to-build-a-lung-disease-prediction-
model-using-chest-x-ray-images (accessed on 15 January 2023).
12.
Cohen, J.P.; Bertin, P.; Frappier, V. Chester: A web delivered locally computed chest X-ray disease prediction system. arXiv
2019
,
arXiv:1901.11210.
13.
Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale Chest X-ray Database and Benchmarks
on Weakly-Supervised Classification and Localization of Common Thorax Diseases. arXiv 2017, arXiv:1705.02315.
14.
Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep net-
works via gradient-based localization. In Proceedings of the IEEE international Conference on Computer Vision, Venice, Italy,
22–29 October 2017; pp. 618–626.
15.
Arun, N.; Gaw, N.; Singh, P.; Chang, K.; Aggarwal, M.; Chen, B.; Hoebel, K.; Gupta, S.; Patel, J.; Gidwani, M.; et al. Assessing the
trustworthiness of saliency maps for localizing abnormalities in medical imaging. Radiol. Artif. Intell.
2021
,3, e200267. [CrossRef]
[PubMed]
16.
Ahmed, K.B.; Goldgof, G.M.; Paul, R.; Goldgof, D.B.; Hall, L.O. Discovery of a generalization gap of convolutional neural
networks on COVID-19 X-rays classification. IEEE Access 2021,9, 72970–72979. [CrossRef] [PubMed]
17.
Çallı, E.; Sogancioglu, E.; van Ginneken, B.; van Leeuwen, K.G.; Murphy, K. Deep learning for chest X-ray analysis: A survey.
Med. Image Anal. 2021,72, 102125. [CrossRef] [PubMed]
18.
Jaeger, S.; Candemir, S.; Antani, S.; Wáng, Y.X.J.; Lu, P.X.; Thoma, G. Two public chest X-ray datasets for computer-aided screening
of pulmonary diseases. Quant. Imaging Med. Surg. 2014,4, 475.
19.
Teixeira, L.O.; Pereira, R.M.; Bertolini, D.; Oliveira, L.S.; Nanni, L.; Cavalcanti, G.D.; Costa, Y.M. Impact of lung segmentation on
the diagnosis and explanation of COVID-19 in chest X-ray images. Sensors 2021,21, 7116. [CrossRef]
20.
de Sousa Freire, N.; de Souza Leão, P.P.; Tiago, L.A.; Gonçalves, A.d.A.C.; Pinto, R.A.; dos Santos, E.M.; Souto, E. Generalizability of
CNN on Predicting COVID-19 from Chest X-ray Images. In 2022: Proceedings of the 22nd Brazilian Symposium on Computing Applied to
Health; SBCAS 2022 Companion Proceedings Series; Brazilian Computing Society (SBC): Porto Alegre, Brazil, 2022; pp. 36–47.
21.
Yahyatabar, M.; Jouvet, P.; Cheriet, F. Dense-Unet: A light model for lung fields segmentation in Chest X-ray images. In Proceed-
ings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal,
QC, Canada, 20–24 July 2020; pp. 1242–1245.
22.
Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer
Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [CrossRef]
23.
Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch:
An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019,32, 1–12.
24.
Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. Chexnet:
Radiologist-level pneumonia detection on chest X-rays with deep learning. arXiv 2017, arXiv:1711.05225
25.
Johnson, A.E.; Pollard, T.J.; Berkowitz, S.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.y.; Mark, R.G.; Horng, S. MIMIC-CXR, a
de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 2019,6, 317. [CrossRef]
26.
Towfighi, S.; Agarwal, A.; Mak, D.; Verma, A. Labelling chest X-ray reports using an open-source NLP and ML tool for text data
binary classificatio. medRxiv 2019. [CrossRef]
27.
Oakden-Rayner, L. Exploring the ChestXray14 Dataset: Problems. 2017. Available online: https://laurenoakdenrayner.com/2017
/12/18/the-chestxray14-dataset-problems/ (accessed on 15 January 2023).
28.
Le, T.D.; Noumeir, R.; Rambaud, J.; Sans, G.; Jouvet, P. Detecting of a patient’s condition from clinical narratives using natural
language representation. IEEE Open J. Eng. Med. Biol. 2022,3, 142–149. [CrossRef] [PubMed]
29.
Sauthier, M.; Tuli, G.; Jouvet, P.A.; Brownstein, J.S.; Randolph, A.G. Estimated Pao2: A continuous and noninvasive method to
estimate Pao2 and oxygenation index. Crit. Care Explor. 2021,3, e0546. [CrossRef] [PubMed]
30.
Brossier, D.; El Taani, R.; Sauthier, M.; Roumeliotis, N.; Emeriaud, G.; Jouvet, P. Creating a high-frequency electronic database in
the PICU: The perpetual patient. Pediatr. Crit. Care Med. 2018,19, e189–e198. [CrossRef] [PubMed]
Disclaimer/Publisher’s Note:
The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... In addition, they utilized GRAD-Cam to highlight the potential ARDS findings on CXRs through saliency maps. On the other hand, Yahyatabar et al. [11] developed the Dense-Ynet model for stratifying the severity of ARDS in CXR images by performing the segmentation and classification tasks simultaneously. A global ARDS severity score for the CXRs was provided based on the distribution of infiltrates in different lung quadrants. ...
... This is particularly notable given the substantial inter-reviewer variability and poor agreements in ARDS diagnosis. For example, Yahyatabar et al. [11] chose to exclude images with labeling disagreements. Similarly, in studies by Reamaroon et al. [8] and Sjoding et al. [9], although uncertain annotations from multiple clinicians were present in the dataset, the training and validation labels relied solely on mean-aggregated values. ...
Article
Full-text available
Acute Respiratory Distress Syndrome (ARDS) is a life-threatening lung injury for which early diagnosis and evidence-based treatment can improve patient outcomes. Chest X-rays (CXRs) play a crucial role in the identification of ARDS; however, their interpretation can be difficult due to non-specific radiological features, uncertainty in disease staging, and inter-rater variability among clinical experts, thus leading to prominent label noise issues. To address these challenges, this study proposes a novel approach that leverages label uncertainty from multiple annotators to enhance ARDS detection in CXR images. Label uncertainty information is encoded and supplied to the model as privileged information, a form of information exclusively available during the training stage and not during inference. By incorporating the Transfer and Marginalized (TRAM) network and effective knowledge transfer mechanisms, the detection model achieved a mean testing AUROC of 0.850, an AUPRC of 0.868, and an F1 score of 0.797. After removing equivocal testing cases, the model attained an AUROC of 0.973, an AUPRC of 0.971, and an F1 score of 0.921. As a new approach to addressing label noise in medical image analysis, the proposed model has shown superiority compared to the original TRAM, Confusion Estimation, and mean-aggregated label training. The overall findings highlight the effectiveness of the proposed methods in addressing label noise in CXRs for ARDS detection, with potential for use in other medical imaging domains that encounter similar challenges.
... At Sainte-Justine Hospital, several research initiatives have been undertaken in the PICU to develop CDSSs for specific needs, such as assistance in the automated diagnosis of acute respiratory distress syndrome in children based on various physiological and radiological criteria [35,36], assessment of the quality of head injury care in adherence to clinical practice guidelines [37], early detection of ventilator-associated pneumonia [38], and hypoxemia diagnosis and management [39]. Unlike the commercially available CDSSs, these tools developed at the Sainte-Justine Hospital were based on local clinical needs, adapted to patient characteristics in the PICU, and developed in harmony with the existing infrastructure, including devices, data availability, and access. ...
... Decision support systems could be incorporated into this level and connected to a patient's problem. For example, a patient with respiratory failure (a medical problem) could have a CDSS for the early diagnosis of acute respiratory distress syndrome [36] and another for the management of mechanical ventilation if diagnosed with acute respiratory distress syndrome [46]. This level allowed the tracking of significant events, such as procedures performed in the operating room, specific investigations, and consultant visits. ...
Article
Full-text available
Background: Clinical decision-making is a complex cognitive process that relies on the interpretation of a large variety of data from different sources and involves the use of knowledge bases and scientific recommendations. The representation of clinical data plays a key role in the speed and efficiency of its interpretation. In addition, the increasing use of clinical decision support systems (CDSSs) provides assistance to clinicians in their practice, allowing them to improve patient outcomes. In the pediatric intensive care unit (PICU), clinicians must process high volumes of data and deal with ever-growing workloads. As they use multiple systems daily to assess patients’ status and to adjust the health care plan, including electronic health records (EHR), clinical systems (eg, laboratory, imaging and pharmacy), and connected devices (eg, bedside monitors, mechanical ventilators, intravenous pumps, and syringes), clinicians rely mostly on their judgment and ability to trace relevant data for decision-making. In these circumstances, the lack of optimal data structure and adapted visual representation hinder clinician’s cognitive processes and clinical decision-making skills. Objective: In this study, we designed a prototype to optimize the representation of clinical data collected from existing sources (eg, EHR, clinical systems, and devices) via a structure that supports the integration of a home-developed CDSS in the PICU. This study was based on analyzing end user needs and their clinical workflow. Methods: First, we observed clinical activities in a PICU to secure a better understanding of the workflow in terms of staff tasks and their use of EHR on a typical work shift. Second, we conducted interviews with 11 clinicians from different staff categories (eg, intensivists, fellows, nurses, and nurse practitioners) to compile their needs for decision support. Third, we structured the data to design a prototype that illustrates the proposed representation. We used a brain injury care scenario to validate the relevance of integrated data and the utility of main functionalities in a clinical context. Fourth, we held design meetings with 5 clinicians to present, revise, and adapt the prototype to meet their needs. Results: We created a structure with 3 levels of abstraction—unit level, patient level, and system level—to optimize clinical data representation and display for efficient patient assessment and to provide a flexible platform to host the internally developed CDSS. Subsequently, we designed a preliminary prototype based on this structure. Conclusions: The data representation structure allows prioritizing patients via criticality indicators, assessing their conditions using a personalized dashboard, and monitoring their courses based on the evolution of clinical values. Further research is required to define and model the concepts of criticality, problem recognition, and evolution. Furthermore, feasibility tests will be conducted to ensure user satisfaction.
... In 14 studies there was an attempt to develop algorithms based on neural network architectures. The developed models based on neural network architecture such as ResNet-50 (CNN) and Dense-Ynet (DNN) were also tested with promising results such as with Jabbour in 2022 [63] and Yahyataba [71] in 2023. However, when competing with non-neural network models in Yang [40] in 2019, Izadi [62] in 2022, Xu [47] in 2021 and Wang [67] in 2023, neural networks showed no advantage in terms of ROC area under the curve (AUC) or accuracy. ...
Article
Full-text available
Aim Acute respiratory distress syndrome or ARDS is an acute, severe form of respiratory failure characterised by poor oxygenation and bilateral pulmonary infiltrates. Advancements in signal processing and machine learning have led to promising solutions for classification, event detection and predictive models in the management of ARDS. Method In this review, we provide systematic description of different studies in the application of Machine Learning (ML) and artificial intelligence for management, prediction, and classification of ARDS. We searched the following databases: Google Scholar, PubMed, and EBSCO from 2009 to 2023. A total of 243 studies was screened, in which, 52 studies were included for review and analysis. We integrated knowledge of previous work providing the state of art and overview of explainable decision models in machine learning and have identified areas for future research. Results Gradient boosting is the most common and successful method utilised in 12 (23.1%) of the studies. Due to limitation of data size available, neural network and its variation is used by only 8 (15.4%) studies. Whilst all studies used cross validating technique or separated database for validation, only 1 study validated the model with clinician input. Explainability methods were presented in 15 (28.8%) of studies with the most common method is feature importance which used 14 times. Conclusion For databases of 5000 or fewer samples, extreme gradient boosting has the highest probability of success. A large, multi-region, multi centre database is required to reduce bias and take advantage of neural network method. A framework for validating with and explaining ML model to clinicians involved in the management of ARDS would be very helpful for development and deployment of the ML model.
Article
Full-text available
Objectives: We sought to update our 2015 work in the Second Pediatric Acute Lung Injury Consensus Conference (PALICC-2) guidelines for the diagnosis and management of pediatric acute respiratory distress syndrome (PARDS), considering new evidence and topic areas that were not previously addressed. Design: International consensus conference series involving 52 multidisciplinary international content experts in PARDS and four methodology experts from 15 countries, using consensus conference methodology, and implementation science. Setting: Not applicable. Patients: Patients with or at risk for PARDS. Interventions: None. Measurements and main results: Eleven subgroups conducted systematic or scoping reviews addressing 11 topic areas: 1) definition, incidence, and epidemiology; 2) pathobiology, severity, and risk stratification; 3) ventilatory support; 4) pulmonary-specific ancillary treatment; 5) nonpulmonary treatment; 6) monitoring; 7) noninvasive respiratory support; 8) extracorporeal support; 9) morbidity and long-term outcomes; 10) clinical informatics and data science; and 11) resource-limited settings. The search included MEDLINE, EMBASE, and CINAHL Complete (EBSCOhost) and was updated in March 2022. Grading of Recommendations, Assessment, Development, and Evaluation methodology was used to summarize evidence and develop the recommendations, which were discussed and voted on by all PALICC-2 experts. There were 146 recommendations and statements, including: 34 recommendations for clinical practice; 112 consensus-based statements with 18 on PARDS definition, 55 on good practice, seven on policy, and 32 on research. All recommendations and statements had agreement greater than 80%. Conclusions: PALICC-2 recommendations and consensus-based statements should facilitate the implementation and adherence to the best clinical practice in patients with PARDS. These results will also inform the development of future programs of research that are crucially needed to provide stronger evidence to guide the pediatric critical care teams managing these patients.
Article
Full-text available
The rapid progress in clinical data management systems and artificial intelligence approaches enable the era of personalized medicine. Intensive care units (ICUs) are ideal clinical research environments for such development because they collect many clinical data and are highly computerized. Goal: We designed a retrospective clinical study on a prospective ICU database using clinical natural language to help in the early diagnosis of heart failure in critically ill children. Methods: The methodology consisted of empirical experiments of a learning algorithm to learn the hidden interpretation and presentation of the French clinical note data. This study included 1386 patients' clinical notes with 5444 single lines of notes. There were 1941 positive cases (36% of total) and 3503 negative cases classified by two independent physicians using a standardized approach. Results: The multilayer perceptron neural network outperforms other discriminative and generative classifiers. Consequently, the proposed framework yields an overall classification performance with 89% accuracy, 88% recall, and 89% precision. Conclusions: This study successfully applied learning representation and machine learning algorithms to detect heart failure in a single French institution from clinical natural language. Further work is needed to use the same methodology in other languages and institutions.
Article
Full-text available
COVID-19 frequently provokes pneumonia, which can be diagnosed using imaging exams. Chest X-ray (CXR) is often useful because it is cheap, fast, widespread, and uses less radiation. Here, we demonstrate the impact of lung segmentation in COVID-19 identification using CXR images and evaluate which contents of the image influenced the most. Semantic segmentation was performed using a U-Net CNN architecture, and the classification using three CNN architectures (VGG, ResNet, and Inception). Explainable Artificial Intelligence techniques were employed to estimate the impact of segmentation. A three-classes database was composed: lung opacity (pneumonia), COVID-19, and normal. We assessed the impact of creating a CXR image database from different sources, and the COVID-19 generalization from one source to another. The segmentation achieved a Jaccard distance of 0.034 and a Dice coefficient of 0.982. The classification using segmented images achieved an F1-Score of 0.88 for the multi-class setup, and 0.83 for COVID-19 identification. In the cross-dataset scenario, we obtained an F1-Score of 0.74 and an area under the ROC curve of 0.9 for COVID-19 identification using segmented images. Experiments support the conclusion that even after segmentation, there is a strong bias introduced by underlying factors from different sources.
Article
Full-text available
Purpose: To evaluate the trustworthiness of saliency maps for abnormality localization in medical imaging. Materials and methods: Using two large publicly available radiology datasets (Society for Imaging Informatics in Medicine-American College of Radiology Pneumothorax Segmentation dataset and Radiological Society of North America Pneumonia Detection Challenge dataset), the performance of eight commonly used saliency map techniques were quantified in regard to (a) localization utility (segmentation and detection), (b) sensitivity to model weight randomization, (c) repeatability, and (d) reproducibility. Their performances versus baseline methods and localization network architectures were compared, using area under the precision-recall curve (AUPRC) and structural similarity index measure (SSIM) as metrics. Results: All eight saliency map techniques failed at least one of the criteria and were inferior in performance compared with localization networks. For pneumothorax segmentation, the AUPRC ranged from 0.024 to 0.224, while a U-Net achieved a significantly superior AUPRC of 0.404 (P < .005). For pneumonia detection, the AUPRC ranged from 0.160 to 0.519, while a RetinaNet achieved a significantly superior AUPRC of 0.596 (P <.005). Five and two saliency methods (of eight) failed the model randomization test on the segmentation and detection datasets, respectively, suggesting that these methods are not sensitive to changes in model parameters. The repeatability and reproducibility of the majority of the saliency methods were worse than localization networks for both the segmentation and detection datasets. Conclusion: The use of saliency maps in the high-risk domain of medical imaging warrants additional scrutiny and recommend that detection or segmentation models be used if localization is the desired output of the network.Keywords: Technology Assessment, Technical Aspects, Feature Detection, Convolutional Neural Network (CNN) Supplemental material is available for this article. © RSNA, 2021.
Article
Full-text available
Pao2 is the gold standard to assess acute hypoxic respiratory failure, but it is only routinely available by intermittent spot checks, precluding any automatic continuous analysis for bedside tools. Objective: To validate a continuous and noninvasive method to estimate hypoxemia severity for all Spo2 values. Derivation cohort: All patients who had an arterial blood gas and simultaneous continuous noninvasive monitoring from 2011 to 2019 at Boston Children's Hospital (Boston, MA) PICU. Validation cohort: External cohort at Sainte-Justine Hospital PICU (Montreal, QC, Canada) from 2017 to 2020. Prediction model: We estimated the Pao2 using three kinds of neural networks and an empirically optimized mathematical model derived from known physiologic equations. Results: We included 52,879 Pao2 (3,252 patients) in the derivation dataset and 12,047 Pao2 (926 patients) in the validation dataset. The mean function on the last minute before the arterial blood gas had the lowest bias (bias -0.1% validation cohort). A difference greater than or equal to 3% between pulse rate and electrical heart rate decreased the intraclass correlation coefficients (0.75 vs 0.44; p < 0.001) implying measurement noise. Our estimated Pao2 equation had the highest intraclass correlation coefficient (0.38; 95% CI, 0.36-0.39; validation cohort) and outperformed neural networks and existing equations. Using the estimated Pao2 to estimate the oxygenation index showed a significantly better hypoxemia classification (kappa) than oxygenation saturation index for both Spo2 less than or equal to 97% (0.79 vs 0.60; p < 0.001) and Spo2 greater than 97% (0.58 vs 0.52; p < 0.001). Conclusion: The estimated Pao2 using pulse rate and electrical heart rate Spo2 validation allows a continuous and noninvasive estimation of the oxygenation index that is valid for Spo2 less than or equal to 97% and for Spo2 greater than 97%. Display of continuous analysis of estimated Pao2 and estimated oxygenation index may provide decision support to assist with hypoxemia diagnosis and oxygen titration in critically ill patients.
Article
Full-text available
A number of recent papers have shown experimental evidence that suggests it is possible to build highly accurate deep neural network models to detect COVID-19 from chest X-ray images. In this paper, we show that good generalization to unseen sources has not been achieved. Experiments with richer data sets than have previously been used show models have high accuracy on seen sources, but poor accuracy on unseen sources. The reason for the disparity is that the convolutional neural network model, which learns features, can focus on differences in X-ray machines or in positioning within the machines, for example. Any feature that a person would clearly rule out is called a confounding feature. Some of the models were trained on COVID-19 image data taken from publications, which may be different than raw images. Some data sets were of pediatric cases with pneumonia where COVID-19 chest X-rays are almost exclusively from adults, so lung size becomes a spurious feature that can be exploited. In this work, we have eliminated many confounding features by working with as close to raw data as possible. Still, deep learned models may leverage source specific confounders to differentiate COVID-19 from pneumonia preventing generalizing to new data sources (i.e. external sites). Our models have achieved an AUC of 1.00 on seen data sources but in the worst case only scored an AUC of 0.38 on unseen ones. This indicates that such models need further assessment/development before they can be broadly clinically deployed. An example of fine-tuning to improve performance at a new site is given.
Article
Full-text available
Background Acute respiratory distress syndrome (ARDS) is a common, but under-recognised, critical illness syndrome associated with high mortality. An important factor in its under-recognition is the variability in chest radiograph interpretation for ARDS. We sought to train a deep convolutional neural network (CNN) to detect ARDS findings on chest radiographs. Methods CNNs were pretrained on 595 506 radiographs from two centres to identify common chest findings (eg, opacity and effusion), and then trained on 8072 radiographs annotated for ARDS by multiple physicians using various transfer learning approaches. The best performing CNN was tested on chest radiographs in an internal and external cohort, including a subset reviewed by six physicians, including a chest radiologist and physicians trained in intensive care medicine. Chest radiograph data were acquired from four US hospitals. Findings In an internal test set of 1560 chest radiographs from 455 patients with acute hypoxaemic respiratory failure, a CNN could detect ARDS with an area under the receiver operator characteristics curve (AUROC) of 0·92 (95% CI 0·89–0·94). In the subgroup of 413 images reviewed by at least six physicians, its AUROC was 0·93 (95% CI 0·88–0·96), sensitivity 83·0% (95% CI 74·0–91·1), and specificity 88·3% (95% CI 83·1–92·8). Among images with zero of six ARDS annotations (n=155), the median CNN probability was 11%, with six (4%) assigned a probability above 50%. Among images with six of six ARDS annotations (n=27), the median CNN probability was 91%, with two (7%) assigned a probability below 50%. In an external cohort of 958 chest radiographs from 431 patients with sepsis, the AUROC was 0·88 (95% CI 0·85–0·91). When radiographs annotated as equivocal were excluded, the AUROC was 0·93 (0·92–0·95). Interpretation A CNN can be trained to achieve expert physician-level performance in ARDS detection on chest radiographs. Further research is needed to evaluate the use of these algorithms to support real-time identification of ARDS patients to ensure fidelity with evidence-based care or to support ongoing ARDS research. Funding National Institutes of Health, Department of Defense, and Department of Veterans Affairs.
Article
Recent advances in deep learning have led to a promising performance in many medical image analysis tasks. As the most commonly performed radiological exam, chest radiographs are a particularly important modality for which a variety of applications have been researched. The release of multiple, large, publicly available chest X-ray datasets in recent years has encouraged research interest and boosted the number of publications. In this paper, we review all studies using deep learning on chest radiographs published before March 2021, categorizing works by task: image-level prediction (classification and regression), segmentation, localization, image generation and domain adaptation. Detailed descriptions of all publicly available datasets are included and commercial systems in the field are described. A comprehensive discussion of the current state of the art is provided, including caveats on the use of public datasets, the requirements of clinically useful systems and gaps in the current literature.
Article
Rationale: Quantifying ARDS severity is essential for prognostic enrichment to stratify patients for invasive or higher risk treatments, however, the comparative performance of many ARDS severity measures is unknown. Objective: To validate ARDS severity measures for their ability to predict hospital mortality and an ARDS-specific outcome (defined as death from pulmonary dysfunction or the need for extra-corporeal membrane oxygenation [ECMO] therapy). Methods: We compared five individual ARDS severity measures including PaO2/FiO2, oxygenation index, ventilatory ratio, lung compliance, and radiologic assessment of lung edema (RALE); two ARDS composite severity scores including the Murray Lung Injury Score (LIS), and a novel score combining RALE, PaO2/FiO2, and ventilatory ratio; and the APACHE-IV score, using data collected at ARDS onset in patients hospitalized at a single center in 2016 and 2017. Discrimination of hospital mortality and the ARDS specific outcome was evaluated using the area under the receiver operator characteristic curve (AUROC). Measure calibration was also evaluated. Results: Among 340 ARDS patients, 125 (37%) died during hospitalization and 36 (10.6%) had the ARDS-specific outcome, including one who received ECMO. Among the five individual ARDS severity measures, the RALE score had the highest discrimination of the ARDS-specific outcome (AUROC = 0.67, 95% CI 0.58-0.77), although other ARDS severity measures had similar performance. However, their ability to discriminate overall mortality was low. In contrast, the APACHE-IV score best discriminated overall mortality (AUROC = 0.73, 95% CI 0.67-0.79) but was unable to discriminate the ARDS-specific outcome (AUROC = 0.54, 95% CI 0.44-0.65). Among ARDS composite severity scores, the LIS had an AUROC = 0.67 (95% CI, 0.58-0.75) for the ARDS-specific outcome while the novel score had an AUROC = 0.79 (95% CI 0.61-0.79). Patients grouped by quartile of the novel score had an 6%, 2%, 10%, and 24% rate of the ARDS-specific outcome. Conclusion: While most ARDS severity measures had poor discrimination of hospital mortality, they performed better at predicting death from severe pulmonary dysfunction or ECMO needs. A novel composite score had the highest the discrimination of this outcome.
Conference Paper
Automatic and accurate lung segmentation in chest X-ray (CXR) images is fundamental for computer-aided diagnosis systems since the lung is the region of interest in many diseases and also it can reveal useful information by its contours. While deep learning models have reached high performances in the segmentation of anatomical structures, the large number of training parameters is a concern since it increases memory usage and reduces the generalization of the model. To address this, a deep CNN model called Dense-Unet is proposed in which, by dense connectivity between various layers, information flow increases throughout the network. This lets us design a network with significantly fewer parameters while keeping the segmentation robust. To the best of our knowledge, Dense-Unet is the lightest deep model proposed for the segmentation of lung fields in CXR images. The model is evaluated on the JSRT and Montgomery datasets and experiments show that the performance of the proposed model is comparable with state-of-the-art methods.