Skip to main content

Image augmentation and automated measurement of endotracheal-tube-to-carina distance on chest radiographs in intensive care unit using a deep learning model with external validation

Abstract

Background

Chest radiographs are routinely performed in intensive care unit (ICU) to confirm the correct position of an endotracheal tube (ETT) relative to the carina. However, their interpretation is often challenging and requires substantial time and expertise. The aim of this study was to propose an externally validated deep learning model with uncertainty quantification and image segmentation for the automated assessment of ETT placement on ICU chest radiographs.

Methods

The CarinaNet model was constructed by applying transfer learning to the RetinaNet model using an internal dataset of ICU chest radiographs. The accuracy of the model in predicting the position of the ETT tip and carina was externally validated using a dataset of 200 images extracted from the MIMIC-CXR database. Uncertainty quantification was performed using the level of confidence in the ETT–carina distance prediction. Segmentation of the ETT was carried out using edge detection and pixel clustering.

Results

The interrater agreement was 0.18 cm for the ETT tip position, 0.58 cm for the carina position, and 0.60 cm for the ETT–carina distance. The mean absolute error of the model on the external test set was 0.51 cm for the ETT tip position prediction, 0.61 cm for the carina position prediction, and 0.89 cm for the ETT–carina distance prediction. The assessment of ETT placement was improved by complementing the human interpretation of chest radiographs with the CarinaNet model.

Conclusions

The CarinaNet model is an efficient and generalizable deep learning algorithm for the automated assessment of ETT placement on ICU chest radiographs. Uncertainty quantification can bring the attention of intensivists to chest radiographs that require an experienced human interpretation. Image segmentation provides intensivists with chest radiographs that are quickly interpretable and allows them to immediately assess the validity of model predictions. The CarinaNet model is ready to be evaluated in clinical studies.

Graphical Abstract

Background

Chest radiographs in the intensive care unit (ICU) are routinely performed to search for signs of pneumonia and pleural effusion or to check the placement of gastric tubes, central catheters, and endotracheal tubes (ETT) [1,2,3].

In the ICU, chest radiographs are usually performed with the patient in a semi-sitting or supine position. As a result, they are often of inferior quality to those performed in the radiology department, and their interpretation typically requires substantial time and expertise [4]. In this context, the automated interpretation of chest radiographs using artificial intelligence could help save time and limit human error by focusing the attention of intensivists on malpositioned ETTs. Artificial intelligence algorithms have shown excellent performance in medical imaging, with deep learning algorithms proving especially effective for the interpretation of fundus images [5, 6], brain magnetic resonance imaging [7], cardiovascular imaging [8], and chest radiographs [9]. The good performance of artificial intelligence algorithms for the assessment of ETT placement on chest radiographs has also been demonstrated. Some studies have successfully used template matching and shape recognition of anatomical landmarks for the detection of the ETT tip and carina [10, 11]. Others have shown the effectiveness of deep learning algorithms trained for the segmentation of the ETT [12], the detection of the ETT tip and carina [13], or the prediction of the ETT–carina distance [14]. However, none of these algorithms includes uncertainty quantification, and none combines ETT tip and carina detection with ETT segmentation. This is unfortunate, as uncertainty quantification would increase the reliability of model predictions, while ETT segmentation would improve the interpretability of chest radiographs in real-world use.

The aim of this study was to propose an externally validated deep learning model with uncertainty quantification and image segmentation for the automated assessment of ETT placement on unselected ICU chest radiographs.

Methods

Study design

The CarinaNet model was constructed by applying transfer learning to the RetinaNet model using an internal dataset composed exclusively of unselected ICU chest radiographs. This dataset called RadioICU was created especially for this work.

The accuracy of the CarinaNet model in predicting the position of the ETT tip and carina on ICU chest radiographs was externally validated using a dataset extracted from the MIMIC-CXR database, a US multicenter database of chest radiographs [15].

Uncertainty quantification was performed using the level of confidence in the ETT–carina distance prediction. Segmentation of the ETT was carried out using edge detection and pixel clustering. In accordance with the Consort-AI guidelines [16], the model code and weights are available at https://github.com/USM-CHU-FGuyon/CarinaNet .

Radio ICU dataset

The RadioICU dataset was composed of all consecutive chest radiographs performed in the polyvalent ICU of Reunion Island University Hospital between 01/01/2016 and 12/31/2019. All radiographs were acquired using a Siemens (München, Germany) Mobilett XP Hybrid Analog Imager with Fuji (Tokyo, Japan) radio luminescent memory screens.

Chest radiographs showing non-intubated patients were retained. Those showing intubation through tracheostomy were dropped from the dataset.

Radiographs were converted from their original DICOM format to 8-bit PNG images and the PixelSpacing attribute was extracted for the conversion from pixel to centimeters. The mean resolution was 3.0 ± 0.5 megapixels. The images were anonymized and did not contain burned-in text.

Images were independently annotated by two experienced intensivists (JA or AR), over a predetermined 3-day period. Intensivists were blinded to each other. The annotation of images included the following information:

  • A subjective quality score between 0 = barely readable; 1 = hardly visible carina; 2 = poor quality; 3 = acceptable quality;

  • The pixel coordinates of the ETT tip;

  • The pixel coordinates of the carina;

  • A classification of the ETT tip relative to the carina : good; too high; too low; bronchial insertion

No coordinates were provided when the carina or the ETT tip could not be located. The annotated pixel coordinates were used as ground truth positions of the ETT tip and carina. A 2-cm distance is commonly cited in the literature as a minimal threshold for the positioning of the ETT relative to the carina [2, 17]. Hence, the ETT position was defined as too low when the ground truth ETT–carina distance was lower than 2cm. The intensivists were asked to classify the too low ETTs following their clinical practices, and were informed that the 2-cm threshold would be used as ground truth for the identification of too low ETTs. After classification, they annotated precisely the pixel coordinates of the ETT tip and carina.

Among radiographs where both the ETT tip and carina coordinates were provided, 200 were randomly selected. This constituted the internal test. No radiographs were excluded on criteria of quality or readability.

MIMIC-CXR dataset

The CarinaNet model was externally validated using a dataset extracted from the MIMIC-CXR database, a US multicenter database of chest radiographs. This external test set allowed to evaluate the generalizability of the CarinaNet model [15].

The MIMIC-CXR dataset was created by randomly selecting 200 chest radiographs featuring an ETT from the MIMIC-CXR database. Radiographs were selected without consideration for image quality or medical report content. Unlike the RadioICU dataset, the MIMIC-CXR dataset contained images with burned-in text. The MIMIC-CXR dataset had higher resolution than the RadioICU dataset images.

On a given day (\(t_0\)), the chest radiographs of the MIMIC-CXR dataset were independently annotated by two experienced intensivists (NA and JA). Ten days later (\(t_{10}\)), the chest radiographs of the MIMIC-CXR dataset were shuffled and given back to one of the two intensivists (JA) for a second blinded annotation. The annotations at \(t_0\) and \(t_{10}\) consisted of the pixel coordinates of the ETT tip and carina.

The ground truth positions of the ETT tip and carina were defined by averaging the two annotations made at \(t_0\). The interrater agreement was defined as the mean absolute difference between the two annotations made at \(t_0\). The intrarater agreement was defined as the mean absolute difference between the annotations made by the same intensivist (JA) at \(t_0\) and \(t_{10}\). Finally, the ETT position was defined as too low when the ground truth ETT–carina distance was lower than 2cm.

The CarinaNet model: a deep learning model for the automated assessment of endotracheal tube placement on chest radiographs

The CarinaNet model was constructed by applying transfer learning to the RetinaNet model, an object detection model developed by Facebook AI Research (New York, USA) [18]. The RetinaNet model has shown state-of-the-art performance on the Common Objects in Context dataset which consists in the detection of specific objects in photographs of everyday life [19]. It has also been used with great success in aerial imagery [20, 21], and medical imaging (computed tomography scans [22], ultrasounds [23], chest radiographs [24], etc.).

The RetinaNet architecture combines a ResNet architecture with a Feature Pyramid Network that outputs a series of feature maps at different spatial scales [25]. These feature maps are fed to two subnetworks: 1) a box subnetwork that builds an anchor box around the object; and 2) a classification subnetwork that classifies the anchor boxes (Fig. 1).

Fig. 1
figure 1

Architecture of the RetinaNet model

The CarinaNet model was constructed by training the RetinaNet model on the RadioICU chest radiographs. For this purpose, clinician’s annotations were converted to (\(200 \times 200\) pixel) bounding boxes around the pixel coordinates of the ETT tip and carina, as shown in Fig. 2.

Fig. 2
figure 2

Bounding boxes around the carina and the endotracheal tube tip on a chest radiograph

For each image, the model output a bounding box for the ETT tip and carina each classified with a confidence score. The position predictions were defined as the center of each bounding box.

Model performance assessment

We evaluated the model performance by computing the mean absolute error between the ETT tip position prediction, the carina position prediction, the ETT–carina distance prediction and their respective ground truths.

The mean errors between the prediction and ground truth for the ETT tip position, carina position, and ETT–carina distance were computed using a linear mixed model. Correlation between measures was accounted for by using a compound symmetry variance–covariance matrix. The same method was used for assessing the interrater and intrarater mean errors. All statistical tests were performed at a two-tailed type I error of 5%.

Uncertainty quantification

To increase the reliability of the CarinaNet model, uncertainty quantification was performed using the level of confidence in the ETT–carina distance prediction.

The level of confidence in the ETT–carina distance prediction was defined as the lowest confidence score between the ETT tip position prediction and the carina position prediction. Expression (1) was used to give the uncertainty U of the ETT–carina distance prediction from the level of confidence C.

$$\begin{aligned} U = \alpha + \beta \ \mathrm{e}^{-\gamma C} \end{aligned}$$
(1)

The ETT tip position was classified as follows :

$$\begin{aligned} {\left\{ \begin{array}{ll} \text {Good position},&{} \text {if } D_{pred} - U_{pred} \ge 2cm \\ \text {Too low}, &{} \text {if } D_{pred} - U_{pred} < 2cm \end{array}\right. }\text {,} \end{aligned}$$
(2)

where \(D_{pred}\) is the ETT–carina distance prediction and \(U_{pred}\) is the uncertainty of the ETT–carina distance prediction.

The use of uncertainty quantification allowed for a higher sensitivity in the classification of too low ETTs. Chest radiographs showing a bronchial insertion were classified along with radiographs showing a too low position of the ETT tip.

Lastly, the accuracy in classifying the ETT tip position as good or too low was compared between a clinician, the CarinaNet model, and a coupled reading. In the latter case, the ETT tip position was classified as too low when either the model or the clinician had classified it as such.

Image augmentation

For each chest radiograph, the following operations were performed to identify the full ETT.

First, a 6-cm-wide rectangular region of interest was selected that extended from 1 cm below the prediction to the top of the image. The resulting smaller image contained the entire ETT.

Second, ridge detection was performed by computing the eigenvalues of the hessian matrix of the image [26], knowing that in image processing a ridge denotes a bright curve against a darker background. This operation resulted in an edge map, containing only the brighter and sharper elements of the chest radiograph.

Third, all non-vertical elements were removed from the edge map using vertical filter (3) on the assumption that the ETT was nearly vertical on all chest radiographs. Morphological opening was applied to the edge map for noise reduction. A binary edge map was obtained by setting the first decile pixel values to 1 and the rest to 0. This binary image was treated as a 2D point cloud.

$$\begin{aligned} \begin{bmatrix} 1 &{} 0 &{} -1\\ \vdots &{} \vdots &{} \vdots \\ 1 &{} 0 &{} -1 \end{bmatrix}_{10\times 3} \end{aligned}$$
(3)

Fourth, clustering of the binary edge map was performed using DBSCAN (Density Based Spatial Clustering of Applications with Noise) [27]. This algorithm clusters point clouds based on the density of neighboring points.

Fifth, the ETT was segmented using a cluster thinning algorithm developed specifically for this study and is illustrated in Fig. 3.

Fig. 3
figure 3

Illustration of the cluster thinning algorithm

This algorithm selected the cluster located closest to the CarinaNet prediction of the ETT tip. It started with the point of the cluster located closest the CarinaNet prediction and then iteratively propagated upwards along the cluster.

Finally, the model output a list of points following a single ridge, which constituted the ETT segmentation.

Results

Annotation of chest radiographs

RadioICU dataset

The RadioICU dataset was composed of 1,890 separate radiographs, 1,357 of which had annotations for both the ETT and carina. The clinicians rated 12.9% of the radiographs as acceptable quality, 70.1% as poor quality, 15.6% as hardly visible carina, and 1.3% as barely readable.

Clinicians generally classified the position of the ETT tip as too low when the ETT–carina distance was less than 2 cm (Fig. 4).

Fig. 4
figure 4

Number of chest radiographs per classification of the endotracheal tube tip position as a function of the annotated endotracheal tube–carina distance

Mimic-CXR dataset

The interrater agreement and intrarater agreement obtained from the annotation of 200 chest radiographs are given in Table 1. The intrarater agreement was slightly lower than the interrater agreement. The latter was used as the reference for human accuracy in interpreting chest radiographs in clinical practice. Table 1 indicates that the readings of the ETT tip and carina positions from the same intensivist (JA) at a 10-day interval were not statistically different, whereas the readings from two separate intensivists (NA and JA) were statistically different.

Predictive performance of the CarinaNet model

The mean absolute error of the CarinaNet model for the ETT tip position prediction, the carina position prediction, and the ETT–carina distance prediction performed on the internal and external test sets are shown in Table 1. The mean absolute error on radiographs from the external test set where the level of confidence was superior to 0.5 was also computed. The CarinaNet mean error for the prediction of the ETT tip position, carina position, and ETT–carina distance indicated that the model predictions were not statistically different from the ground truth positions.

Table 1 Model performance, interrater and intrarater agreements for the position of the endotracheal tube (ETT) tip, carina, and ETT–carina distance

Table 2 shows the model’s performance on the RadioICU dataset for each radiograph quality. The best predictive performance was obtained with the more frequent radiograph qualities (poor quality, hardly visible carina) rather than with the acceptable quality radiographs.

Table 2 Mean absolute error on all images of the RadioICU dataset where coordinates were provided for both the endotracheal tube and carina. ETT : endotracheal tube

Uncertainty quantification

Figure 5 indicates an exponential decrease in the absolute error of the ETT–carina distance prediction with the increase in the level of confidence. This justified the use of equation (1).

Fig. 5
figure 5

Exponential fit of the absolute error for the endotracheal tube–carina distance prediction as a function of the level of confidence in this prediction. The predictions were grouped into 8 bins based on the level of confidence and the mean absolute error was computed for each bin

Based on the exponential fit shown in Fig. 5, the uncertainty quantification of the CarinaNet model was defined as:

$$\begin{aligned} U_{pred} = {0.1} + {4.2}\ \mathrm{e}^{{-3.3}C} \end{aligned}$$
(4)

where \(U_{pred}\) is the uncertainty of the ETT–carina distance prediction and C is the level of confidence in the ETT–carina distance prediction.

Confusion matrices for the classification of the ETT tip position by a clinician, the CarinaNet model, and a coupled reading are shown in Fig. 6. The coupled reading classified the ETT tip as too low, when either the CarinaNet model or the clinician classified it as such. In each matrix, the classification of the ETT tip position was compared to a measure of the effective ETT–carina distance. As Fig. 6 indicates, 8 chest radiographs were wrongly classified as showing a well-positioned ETT tip when the classification was performed by a clinician. This number fell to 3 when the classification was performed by a clinician together with the CarinaNet model.

Fig. 6
figure 6

Confusion matrices for the classification of the endotracheal tube tip position by a clinician, the CarinaNet model, and a coupled reading

The sensitivity and specificity of the classification of too low ETT tips by a clinician, the CarinaNet model, and a coupled reading are shown in Table 3. The lower specificity of the CarinaNet model was expected when using classification criterium (2), but induced a better sensitivity in the context of a combination with the human interpretation.

Table 3 Sensitivity and specificity of the classification of low endotracheal tip location by a clinician, the CarinaNet model, and a coupled reading

Image segmentation

The successive steps of the image segmentation performed on a chest radiograph of the RadioICU dataset are shown in Figs. 7 and 8 .

First, a region of interest centered on the ETT was obtained thanks to the good accuracy of the ETT tip position prediction (Fig. 7).

Fig. 7
figure 7

Edge detection on the region of interest

Fig. 8
figure 8

Clustering and cluster thinning on the binary edge map

Second, the ridge detection algorithm was applied to the chest radiograph. The resulting edge map clearly highlighted the ETT. However, other elements of the region of interest were also highlighted, in particular the horizontal cable at the bottom of the image and the small artifacts on either side of the ETT (Fig. 7).

Third, all non-vertical elements and smaller artifacts were removed from the original image by applying the vertical filter (3) to the edge map and then performing morphological opening. The horizontal cable and the smaller edges totally disappeared, leaving the entire ETT clearly highlighted (Fig. 7). After binarization of the filtered edge map, the only remaining elements were the ETT, the gastric tube, and certain artifacts (Fig. 8).

Fourth, non-overlapping objects were separated by applying the DBSCAN algorithm to the binary edge map. The artifacts located on either side of the ETT were classified into separate clusters. As the gastric tube considerably overlapped with the ETT on the original chest radiograph, the DBSCAN algorithm generated a single cluster containing both objects (Fig. 8).

Fifth, the cluster thinning algorithm was applied to the binary edge map. This operation highlighted the ETT but not the gastric tube (Fig. 8).

The output of the cluster thinning algorithm constituted the ETT segmentation. The segmentation obtained from the successive operations shown in Figs. 7 and 8 is visible on the top-left radiograph in Fig. 9. As Fig. 9 indicates, the CarinaNet model predicted the ETT–carina distance with very low uncertainty. Bottom-left radiograph of Fig. 9 shows the segmentation of the ETT on a tilted radiograph. On this radiograph, the ETT was entirely segmented, but the segmentation also selected the border of the image.

Fig. 9
figure 9

Example outputs of the CarinaNet model. Top-left and top-right chest radiograph were, respectively, classified as having acceptable quality and a hardly visible carina. The endotracheal tube tip and carina were accurately detected and the model uncertainty was inferior to 1 cm, indicating that the model result could be trusted. The segmentation was successful for both these radiographs. Bottom-left radiograph had a very unusual angle, this disrupted the segmentation step. However the endotracheal tube was entirely segmented and the endotracheal tube tip and carina were accurately detected. Bottom-right radiograph featured no endotracheal tube. The CarinaNet model properly detected the carina and output an endotracheal tube tip position but with an uncertainty of 3.3 cm indicating that the result could not be trusted. Note that the segmentation step also failed

Discussion

This is the first study to propose an externally validated deep learning algorithm with uncertainty quantification and image segmentation for the automated assessment of ETT placement on unselected ICU chest radiographs. The main benefit of the CarinaNet model is that it can help intensivists prioritize chest radiographs that have been pre-classified as featuring a too low ETT. Uncertainty quantification is also a major security contribution in that it can bring the attention of intensivists to chest radiographs that require an experienced human interpretation. Lastly, the usability of the CarinaNet model is reinforced by image segmentation, which provides intensivists with chest radiographs that are quickly interpretable and allows them to immediately assess the validity of model predictions.

Little difference was observed between the internal validation of the CarinaNet model and its external validation. Indeed, the mean absolute error of the model for the ETT tip position, carina position and ETT–carina distance was 0.51 cm, 0.61 cm, 0.89 cm on the external test set compared to 0.59 cm, 0.60 cm, 0.90 cm on the internal test set, respectively.

This good generalization of the CarinaNet model indicates that it is not disturbed by unfamiliar elements on chest radiographs. More generally, the good predictive performance of the model suggests that it is usable in clinical practice.

Previous studies have shown the interest of using deep learning algorithms for the automated interpretation of ICU chest radiographs. In a study evaluating the InceptionV3 model, external validation on 100 images yielded an error of 0.63 ± 0.55 cm for the ETT–carina distance prediction [14]. This suggests a better performance than the CarinaNet model; however, the model has low interpretability as it did not explicitly detect the ETT tip and carina. Another study evaluated a cascaded convolutional neural network model for its accuracy in predicting the ETT tip or carina position on ICU chest radiographs from the MIMIC-CXR database [13]. The prediction error of this model on the internal test set was 0.82, 0.64 and 0.86 cm for the ETT tip position, carina position, and ETT–carina distance, respectively [13]. While these findings are similar to ours, it should be noted that this study proposed no external validation. Moreover, in both these studies the models were evaluated on radiographs for which the ETT–carina distance was already specified in the medical report. This likely induced a selection bias.

The CarinaNet model was able to identify too low ETT tip positions with a sensitivity of 0.85 and a specificity of 0.87. The sensitivity and specificity of the classification performed by a clinician were 0.76 and 0.99, respectively, while those of the classification performed by a clinician coupled with the CarinaNet model were 0.91 and 0.87, respectively. This increase in sensitivity corresponds to increased security compared to current clinical practice. The lower specificity of the CarinaNet model indicates that pre-classification may prioritize the clinician’s reading of some chest radiographs showing a well-positioned ETT. Thus, this pre-classification remains a strict improvement from current clinical practice. Our findings support studies that recommend using artificial intelligence as a complement to human tasks rather than as a substitute [28].

There was a considerable difference between the prediction error of the CarinaNet model and the interrater agreement for the ETT tip position. This can be explained by the fact that the ETT was designed to be radiopaque and easily identifiable by the human eye. By contrast, the prediction error was very close to the interrater agreement for the carina position. The model performance even surpassed the interrater agreement on the carina detection when considering predictions where the level of confidence was superior to 0.5, which was half of the dataset. These findings suggest that deep learning algorithms have different perceptive qualities than the human eye and can therefore be used to complement the human interpretation of chest radiographs.

Additionally, the model predictions were not statistically different from the ground truth for the ETT tip and carina positions. By contrast, the readings of the ETT tip and carina positions by two experienced intensivists were statistically different. This demonstrates that our CarinaNet model could improve the current clinical practices by reducing the impact of the inter-intensivist variability for the reading of the routine chest radiographs.

Interestingly, the lowest prediction errors of the CarinaNet model were obtained not for chest radiographs classified by intensivists as being of acceptable quality, but for those of lower quality, which were the most frequent. This shows that deep learning models trained with hard examples have good adaptability to challenging tasks. The CarinaNet model yielded the highest prediction errors for images taken from a very unusual angle (see Fig. 9) or featuring opaque lungs. This problem could be addressed by increasing the proportion of tilted images or images of opaque lungs in the training set.

Moreover, we designed our model to be used solely on chest radiographs of intubated patients because we believe the pre-interpretation should only be made when clinically relevant. As a consequence, our model did not explicitly check for the presence of an endotracheal tube. However, our built-in uncertainty quantification acted as a safeguard by indicating that the result could not be trusted when the uncertainty was too high (bottom-right radiograph of Fig. 9). This demonstrates that integrated uncertainty quantification is a major robustness contribution for the clinical application of our model.

Our study has some limitations. First, since the images of the training set were non-redundantly annotated, the prediction error of the model was necessarily higher than the interrater agreement for the carina position. Second, few of the test set images showed a malpositioned ETT, which induced imprecise sensitivity and specificity values for the model’s classification. This problem could be overcome by using test sets with a larger proportion of images showing a malpositioned ETT. Third, the coupled reading was simulated using the model’s classification and the clinician’s blinded annotation. Our results thus underestimate the benefit of using the CarinaNet model as the clinician’s reading should be evaluated with the knowledge of the model’s classification as well as the augmented image. Finally, the use of the 2-cm criterion could be debated as no study provided evidence for the superiority of this threshold against another. However, our model is adaptable to any given threshold if recommendations change in the future.

In the future, the predictive performance of the CarinaNet model could be increased by using a training set with a larger proportion of images showing a malpositioned ETT. Data augmentation in training or at test time could also improve the performance of the model. Lastly, studies should be performed to validate the applicability of the model in clinical practice.

Conclusion

In conclusion, the CarinaNet model is an efficient and generalizable deep learning algorithm for the automated assessment of ETT placement on ICU chest radiographs. This model can help intensivists prioritize chest radiographs that have been pre-classified as featuring a malpositioned ETT. Moreover, uncertainty quantification can bring the attention of intensivists to chest radiographs that require an experienced human interpretation. Finally, image segmentation provides intensivists with chest radiographs that are quickly interpretable and allows them to immediately assess the validity of model predictions. The CarinaNet model is ready to be evaluated in clinical trials.

Availability of data and materials

The RadioICU dataset—used for training and internal validation—is not publicly available. The MIMIC-CXR dataset—used for external validation—is publicly available at https://physionet.org/content/mimic-cxr/2.0.0/. All the codes and model weights needed to reproduce the results are available at https://github.com/USM-CHU-FGuyon/CarinaNet.

Abbreviations

ICU:

Intensive care unit

ETT:

Endotracheal tube

References

  1. Hejblum G, Chalumeau-Lemoine L, Ioos V, et al. Comparison of routine and on-demand prescription of chest radiographs in mechanically ventilated adults: a multicentre, cluster-randomised, two-period crossover study. Lancet. 2009;374:1687–93.

    Article  Google Scholar 

  2. Colin C. Prise en charge des voies aériennes en anesthésie adulte à l’exception de l’intubation difficile. Annales Françaises d’Anesthésie et de Réanimation. 2003;22:3–17.

    Article  Google Scholar 

  3. Higgs A, Mcgrath B, Goddard C, et al. Guidelines for the management of tracheal intubation in critically ill adults. Br J Anaesth. 2017;120:323–52.

    Article  Google Scholar 

  4. Ganapathy A, Adhikari N, Spiegelman J, Scales D. Routine chest x-rays in intensive care units: a systemic review and meta-analysis. Crit Care. 2012;16:R68.

    Article  Google Scholar 

  5. Fauw J, Ledsam J, Romera-Paredes B, et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat Med. 2018;24:1342–50.

    Article  Google Scholar 

  6. Gulshan V, Peng L, Coram M, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316:2402–10.

    Article  Google Scholar 

  7. Akkus Z, Galimzianova A, Hoogi A, Rubin D, Erickson B. Deep learning for brain MRI segmentation: state of the art and future directions. J Digit Imaging. 2017;30:449–59.

    Article  Google Scholar 

  8. Savadjiev P, Chong J, Dohan A, et al. Demystification of AI-driven medical image interpretation: past, present and future. Eur Radiol. 2018;29:1616–24.

    Article  Google Scholar 

  9. Lee H, Mansouri M, Tajmir S, Lev M, Do S. A deep-learning system for fully-automated peripherally inserted central catheter (PICC) tip detection. J Digit Imaging. 2017;31:1–10.

    Google Scholar 

  10. Zhimin H, Jing Z, inventors; Carestream Health Inc, assignee. Method for detecting anatomical structures. US 8,577,108 B2; 2013.

  11. Laoa Z, Zhengb X, Zouc Q. Carina landmark detection in ICU images via integrating geometrical and thoracic anatomy based features. In: 21st international conference on pattern recognition. 2012; p. 5–8.

  12. Frid-Adar M, Amer R, Greenspan H. Medical image computing and computer assisted intervention. In: Endotracheal tube detection and segmentation in chest radiographs using synthetic data. Cham: Springer; 2019. p. 784–92.

    Google Scholar 

  13. Kara S, Akers J, Chang P. Identification and Localization of endotracheal tube on chest radiographs using a cascaded convolutional neural network approach. J Digit Imaging. 2021;34:898–904.

    Article  Google Scholar 

  14. Lakhani P, Flanders A, Gorniak R. Endotracheal tube position assessment on chest radiographs using deep learning. Radiol Artif Intell. 2021;3: e200026.

    Article  Google Scholar 

  15. Johnson A, Pollard T, Berkowitz S, et al. MIMIC-CXR: a large publicly available database of labeled chest radiographs. Scientific Data. 2019;6.

  16. Liu X, Cruz Rivera S, Moher D, et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: The CONSORT-AI Extension. Nat Med. 2020;26:1364–74.

    Article  CAS  Google Scholar 

  17. Geisser W, Maybauer D, Wolff H, Pfenninger E, Maybauer M. Radiological validation of tracheal tube insertion depth in out-hospital and in-hospital emergency patients. Anaesthesia. 2009;64:973–7.

    Article  CAS  Google Scholar 

  18. Lin TY, Goyal P, Girshick R, He K, Dollar P. Focal Loss for Dense Object Detection. IEEE International Conference on Computer Vision. 2017; p. 2999–3007.

  19. Lin TY, Maire M, Belongie S, et al. Microsoft COCO: common objects in context. European conference on computer vision. 2014;05:740–55.

    Google Scholar 

  20. Wang Y, Wang C, Zhang H, Dong Y, Wei S. Automatic ship detection based on RetinaNet using multi-resolution Gaofen-3 imagery. Remote Sens. 2019;11:531.

    Article  CAS  Google Scholar 

  21. Wei H, Zhao Y, Dong J. Cooling tower detection based on the improved RetinaNet. Remote Sens Nat Resour. 2020;32:68–73.

    Google Scholar 

  22. Zlocha M, Dou Q, Glocker B; Springer. Improving RetinaNet for CT lesion detection with dense masks from weak RECIST labels. In: International conference on medical image computing and computer-assisted intervention. 2019; p. 402–10.

  23. Yang M, Xiao X, Liu Z, et al. Deep RetinaNet for dynamic left ventricle detection in multiview echocardiography classification. Sci Program. 2020;2020:1–6.

    CAS  Google Scholar 

  24. Mao L, Yumeng T, Lina C. Pneumonia detection in chest X-rays: a deep learning approach based on ensemble RetinaNet and Mask R-CNN. In: Eighth international conference on advanced Cloud and Big Data (CBD). 2020; p. 213–18.

  25. Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S. Feature Pyramid Networks for Object Detection. In: IEEE conference on computer vision and pattern recognition. 2017; p. 936–44.

  26. Staal J, Abramoff MD, Niemeijer M, Viergever MA, van Ginneken B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans Med Imaging. 2004;23(4):501–9.

    Article  Google Scholar 

  27. Ester M, Kriegel HP, Sander J, Xu X. A density-based algorithm for discovering clusters in large spatial databases with noise. In: Proceedings of the second international conference on knowledge discovery and data mining. 1996; p. 226–31.

  28. Gruber K. Is the future of medical diagnosis in computer algorithms? Lancet Digit Health. 2019;05(1):e15–6.

    Article  Google Scholar 

Download references

Acknowledgements

We thank Arianne Dorval for her editorial assistance.

Funding

None.

Author information

Authors and Affiliations

Authors

Contributions

MO did the formal analysis, investigation, data curation, methodology, software, visualization and wrote the original draft. JA, NA, CF, and AR did the conceptualization, data curation, funding acquisition, investigation, project administration, resources, supervision, validation, and reviewed the manuscript. LM did the methodology, data curation, resources, and reviewed the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Matthieu Oliver.

Ethics declarations

Ethics approval and consent to participate

This study was conducted in accordance with the regulations of the French Commission on Information Technology and Liberties (CNIL) and was approved by the Ethics Committee of the French Intensive Care Society (#CE-SRLF-20-69). The data were anonymized and de-identified. No data were collected other than anonymized images.

Consent for publication

Not applicable. The individual data presented in this article were anonymized and not identifiable.

Competing interest

The authors declare that they have no competing interests

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Oliver, M., Renou, A., Allou, N. et al. Image augmentation and automated measurement of endotracheal-tube-to-carina distance on chest radiographs in intensive care unit using a deep learning model with external validation. Crit Care 27, 40 (2023). https://doi.org/10.1186/s13054-023-04320-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13054-023-04320-0

Keywords