Deep Convolutional Neural Networks for Endotracheal Tube Position and X-ray Image Classification: Challenges and Opportunities

Paras Lakhani, Paras Lakhani

Abstract

The goal of this study is to evaluate the efficacy of deep convolutional neural networks (DCNNs) in differentiating subtle, intermediate, and more obvious image differences in radiography. Three different datasets were created, which included presence/absence of the endotracheal (ET) tube (n = 300), low/normal position of the ET tube (n = 300), and chest/abdominal radiographs (n = 120). The datasets were split into training, validation, and test. Both untrained and pre-trained deep neural networks were employed, including AlexNet and GoogLeNet classifiers, using the Caffe framework. Data augmentation was performed for the presence/absence and low/normal ET tube datasets. Receiver operating characteristic (ROC), area under the curves (AUC), and 95% confidence intervals were calculated. Statistical differences of the AUCs were determined using a non-parametric approach. The pre-trained AlexNet and GoogLeNet classifiers had perfect accuracy (AUC 1.00) in differentiating chest vs. abdominal radiographs, using only 45 training cases. For more difficult datasets, including the presence/absence and low/normal position endotracheal tubes, more training cases, pre-trained networks, and data-augmentation approaches were helpful to increase accuracy. The best-performing network for classifying presence vs. absence of an ET tube was still very accurate with an AUC of 0.99. However, for the most difficult dataset, such as low vs. normal position of the endotracheal tube, DCNNs did not perform as well, but achieved a reasonable AUC of 0.81.

Keywords: Artificial intelligence; Artificial neural networks (ANNs); Classification; Machine learning; Radiography.

Figures

Fig. 1
Fig. 1
ROC curves of the best-performing classifiers for the three different datasets. The AUC of the chest/abdomen model was the highest at 1.000 (solid black line), followed by the presence/absence at 0.989 (dashed black line), and low/normal at 0.809 (dotted black line)
Fig. 2
Fig. 2
With reduced learning rate of 0.001, there is fast convergence using pre-trained weights on the AlexNet model. Both training and validation loss reduce over the course of training, with accuracy on the validation dataset reaching approximately 99% at 90 epochs
Fig. 3
Fig. 3
With higher learning rate of 0.01, there is no convergence after 90 epochs using pre-trained weights on the AlexNet model, likely due to large gradients that impact the learning process. The training and validation loss remains high with accuracy constant at 50%
Fig. 4
Fig. 4
Effect of number of training cases on AUCs for the pre-trained models for the ET presence/absence dataset. Using 100% of training cases resulted in statistically significant higher AUCs than with 25% of the data (P = 0.015 and P < 0.0001, for AlexNet_T and GoogLeNet_T, respectively)
Fig. 5
Fig. 5
Chest radiograph with ET tube present (white arrow)
Fig. 6
Fig. 6
Follow-up chest radiograph on the same patient after ET tube has been removed. The white arrow depicts the absence of the ET tube
Fig. 7
Fig. 7
Cropped chest radiograph image centered on the carina. The ET tube tip is low at the orifice of the right main stem bronchus (white arrow)
Fig. 8
Fig. 8
Cropped chest radiograph image shows that the ET tube tip (white arrow) is in satisfactory position 4 cm above the carina
Fig. 9
Fig. 9
Saliency map on the left with corresponding radiograph on the right, where an ET tube is present. The saliency map highlights parts of the images that contribute most to the prediction class. The black arrows points to the location of the ET tube, which has a visible contribution to the prediction label. However, there are other parts of the image, including edges of ribs and the aortic knob that also contribute to the score for this image. While the network had high accuracy, there is still room for improvement
Fig. 10
Fig. 10
Saliency map on the left with corresponding radiograph on the right with low position of ET tube in the right main stem bronchus. On the saliency map, the black arrows point to the location of the ET tube. The white arrows point to an unrelated overlying catheter and an enteric tube. From the saliency map, the ET tube, enteric tube, and overlying catheter all have contributions to the prediction class, indicating that the network is not as well formed and is inferring from parts of the image (enteric tube and overlying catheter) that are not relevant to the prediction label (ET tube is low). This may explain the lower accuracy of this network compared to the others

References

    1. Goodman LR, Conrardy PA, Laing F, Singer MM: Radiographic evaluation of endotracheal tube position. Am J Roentgenol 127(3):433–4,1976
    1. Brunel W, Coleman DL, Schwartz DE, Peper E, Cohen NH: Assessment of routine chest roentgenograms and the physical examination to confirm endotracheal tube position. Chest 96(5):1043–5,1989
    1. Zwillich CW, Pierson DJ, Creagh CE, Sutton FD, Schatz E, Petty TL: Complications of assisted ventilation: a prospective study of 354 consecutive episodes. Am J Med 57(2):161–70,1974
    1. Ramakrishna B, Brown M, Goldin J, Cagnon C, Enzmann D: An improved automatic computer aided tube detection and labeling system on chest radiographs. Proc. SPIE 8315, Medical Imaging 2012: Computer-Aided Diagnosis, 83150R (February 23, 2012); doi:10.1117/12.911839
    1. Kao EF, Jaw TS, Li CW, Chou MC, Liu GC: Automated detection of endotracheal tubes in paediatric chest radiographs. Comput Methods Prog Biomed 118(1):1–10,2015
    1. Chen S, Zhang M, Yao L, Xu W: Endotracheal tubes positioning detection in adult portable chest radiography for intensive care unit. Int J Comput Assist Radiol Surg 11(11):2049–57,2016
    1. Krizhevsky A, Sutskever I, Hinton GE: Imagenet classification with deep convolutional neural networks. Adv Neural Inf Proces Syst 1097–1105, 2012
    1. Szegedy C, Liu W, Jia Y, et al.: Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition: 1–9, 2015
    1. Russakovsky O, Deng J, Su H, et al.: Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252,2015
    1. Bar Y, Diamant I, Wolf L, Greenspan H: Deep learning with non-medical training used for chest pathology identification. Proc.SPIE Medical Imaging, International Society for Optics and Photonics: 9414, 2015
    1. Rajkomar A, Lingam S, Taylor AG, Blum M, Mongan J. High-throughput classification of radiographs using deep convolutional neural networks. J Digit Imaging. 2016;11:1–7.
    1. Zhang W, Li R, Deng H, Wang L, Lin W, Ji S, Shen D. Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage. 2015;108:214–224. doi: 10.1016/j.neuroimage.2014.12.061.
    1. Shin HC, Roth HR, Gao M, et al.: Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging 35(5):1285–1298,2016
    1. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, Venugopalan S, Widner K, Madams T, Cuadros J, Kim R: Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 316(22):2402–10,2016
    1. Cho J, Lee K, Shin E, Choy G, Do S: How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv preprint arXiv:1511.06348, 2015
    1. Jia Y, Shelhamer E, Donahue J, et al.: Caffe: convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM international conference on Multimedia 2014. ACM, 2014
    1. Hastie T, Tishirani R, Friedman J. Model assessment and selection. In: The Elements of Statistical Learning, 2nd ed. Berlin: Springer; 2009. pp. 219–257.
    1. Steyerberg EW, Vickers AJ, Cook NR et al.: Assessing the performance of prediction models: a framework for some traditional and novel measures. Epidemiology 21(1):128–138,2010
    1. Bradley AP: The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recogn 30(7):1145–1159,1997
    1. Fawcett T.: ROC graphs: notes and practical considerations for researchers. Mach Learn 31(1):1–38,2004
    1. Delrue L, Gosselin R, Ilsen B, Van Landeghem A, de Mey J, Duyck P. Difficulties in the interpretation of chest radiography. In Comparative Interpretation of CT and Standard Radiography of the Chest. Berlin Heidelberg: Springer; 2011. pp. 27–49.
    1. Yosinski J, Clune J, Bengio Y, Lipson H: How transferable are features in deep neural networks? In Advances in neural information processing systems, 2014, pp. 3320–3328
    1. Simonyan K, Vedaldi A, Zisserman A: Deep inside convolutional networks: visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013

Source: PubMed

3
Suscribir