Computer-aided diagnosis of laryngeal cancer via deep learning based on laryngoscopic images

Hao Xiong, Peiliang Lin, Jin-Gang Yu, Jin Ye, Lichao Xiao, Yuan Tao, Zebin Jiang, Wei Lin, Mingyue Liu, Jingjing Xu, Wenjie Hu, Yuewen Lu, Huaifeng Liu, Yuanqing Li, Yiqing Zheng, Haidi Yang, Hao Xiong, Peiliang Lin, Jin-Gang Yu, Jin Ye, Lichao Xiao, Yuan Tao, Zebin Jiang, Wei Lin, Mingyue Liu, Jingjing Xu, Wenjie Hu, Yuewen Lu, Huaifeng Liu, Yuanqing Li, Yiqing Zheng, Haidi Yang

Abstract

Objective: To develop a deep convolutional neural network (DCNN) that can automatically detect laryngeal cancer (LCA) in laryngoscopic images.

Methods: A DCNN-based diagnostic system was constructed and trained using 13,721 laryngoscopic images of LCA, precancerous laryngeal lesions (PRELCA), benign laryngeal tumors (BLT) and normal tissues (NORM) from 2 tertiary hospitals in China, including 2293 from 206 LCA subjects, 1807 from 203 PRELCA subjects, 6448 from 774 BLT subjects and 3191 from 633 NORM subjects. An independent test set of 1176 laryngoscopic images from other 3 tertiary hospitals in China, including 132 from 44 LCA subjects, 129 from 43 PRELCA subjects, 504 from 168 BLT subjects and 411 from 137 NORM subjects, was applied to the constructed DCNN to evaluate its performance against experienced endoscopists.

Results: The DCCN achieved a sensitivity of 0.731, a specificity of 0.922, an AUC of 0.922, and the overall accuracy of 0.867 for detecting LCA and PRELCA among all lesions and normal tissues. When compared to human experts in an independent test set, the DCCN' s performance on detection of LCA and PRELCA achieved a sensitivity of 0.720, a specificity of 0.948, an AUC of 0.953, and the overall accuracy of 0.897, which was comparable to that of an experienced human expert with 10-20 years of work experience. Moreover, the overall accuracy of DCNN for detection of LCA was 0.773, which was also comparable to that of an experienced human expert with 10-20 years of work experience and exceeded the experts with less than 10 years of work experience.

Conclusions: The DCNN has high sensitivity and specificity for automated detection of LCA and PRELCA from BLT and NORM in laryngoscopic images. This novel and effective approach facilitates earlier diagnosis of early LCA, resulting in improved clinical outcomes and reducing the burden of endoscopists.

Conflict of interest statement

The authors declare that they have no competing interests.

Copyright © 2018. Published by Elsevier B.V.

Figures

Fig. 1
Fig. 1
Overview of the deep learning architecture. Parameters pre-trained on the external ImageNet dataset are used to initialize the deep convolutional neural network, which is then fine-tuned on the target dataset.
Fig. 2
Fig. 2
Illustration of the changes of the loss function value (A) and the classification accuracy (B) over the training and testing sets.
Fig. 3
Fig. 3
The sensitivity-specificity curve for Urgent versus Non-urgent binary classification.
Fig. 4
Fig. 4
Confusion matrix for 4-class categorization.
Fig. 5
Fig. 5
Comparison between the deep learning algorithm and three human experts in Urgent versus Non-urgent binary classification. Expert 1: expert with 10–20 years of experience. Expert 2: expert with ~3 years of experience. Expert 3: expert with 3–10 years of experience.
Fig. 6
Fig. 6
Four-class confusion matrices obtained by our DCNN model (A) and the three human experts (B-D). Expert 1: expert with 10–20 years of experience. Expert 2: expert with ~3 years of experience. Expert 3: expert with 3–10 years of experience.
Fig. 7
Fig. 7
Representative attention maps obtained by the DCNN model on the classes of NORM, BLT, PRELCA and LCA from top to bottom respectively. Attention maps are displayed as heat maps overlay upon the original images, where warmer colors indicate higher saliency, i.e., higher contribution to the classification decision.

References

    1. Marioni G., Marchese-Ragona R., Cartei G., Marchese F., Staffieri A. Current opinion in diagnosis and treatment of laryngeal carcinoma. Cancer Treat Rev. 2006;32:504–515.
    1. Ni X.G., Zhang Q.Q., Wang G.Q. Narrow band imaging versus autofluorescence imaging for head and neck squamous cell carcinoma detection: a prospective study. J Laryngol Otol. 2016;130:1001–1006.
    1. Barbalata C., Mattos L.S. Laryngeal tumor detection and classification in endoscopic video. IEEE J Biomed Health Inform. 2016;20:322–332.
    1. Dai Z.H., Chen L.B. Yi chuan = Hereditas. vol. 32. 2010. The impact of microRNAs on the evolution of metazoan complexity; pp. 105–114.
    1. Kraft M., Fostiropoulos K., Gurtler N., Arnoux A., Davaris N., Arens C. Value of narrow band imaging in the early diagnosis of laryngeal cancer. Head Neck. 2016;38:15–20.
    1. De Vito A., Meccariello G., Vicini C. Narrow band imaging as screening test for early detection of laryngeal cancer: a prospective study. Clin Otolaryngol. 2017;42:347–353.
    1. Sun C., Han X., Li X., Zhang Y., Du X. Diagnostic performance of narrow band imaging for laryngeal Cancer: a systematic review and meta-analysis. Otolaryngol Head Neck Surg. 2017;156:589–597.
    1. Yang Y., Liu J., Song F., Zhang S. The clinical diagnostic value of target biopsy using narrow-band imaging endoscopy and accurate laryngeal carcinoma pathologic specimen acquisition. Clin Otolaryngol. 2017;42:38–45.
    1. Esteva A., Kuprel B., Novoa R.A., Ko J., Swetter S.M., Blau H.M. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542:115–118.
    1. Sempere L.F., Cole C.N., McPeek M.A., Peterson K.J. The phylogenetic distribution of metazoan microRNAs: insights into evolutionary complexity and constraint. J Exp Zool B Mol Dev Evol. 2006;306:575–588.
    1. Rose J.M., Novoselov S.S., Robinson P.A., Cheetham M.E. Molecular chaperone-mediated rescue of mitophagy by a parkin RING1 domain mutant. Hum Mol Genet. 2011;20:16–27.
    1. Khosravi P., Kazemi E., Imielinski M., Elemento O., Hajirasouliha I. Deep convolutional neural networks enable discrimination of heterogeneous digital pathology images. EBioMedicine. 2018;27:317–328.
    1. Gulshan V., Peng L., Coram M., Stumpe M.C., Wu D., Narayanaswamy A. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316:2402–2410.
    1. Kermany D.S., Goldbaum M., Cai W., Valentim C.C.S., Liang H., Baxter S.L. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell. 2018;172:1122–1131. [e1129]
    1. Verikas A., Gelzinis A., Valincius D., Bacauskiene M., Uloza V. Multiple feature sets based categorization of laryngeal images. Comput Methods Prog Biomed. 2007;85:257–266.
    1. Hirasawa T., Aoyama K., Tanimoto T., Ishihara S., Shichijo S., Ozawa T. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer. 2018;21:653–660.
    1. Liu Z., Gao J., Yang G., Zhang H., He Y. Localization and classification of paddy field pests using a saliency map and deep convolutional neural network. Sci Rep. 2016;6
    1. Chmelik J., Jakubicek R., Walek P., Jan J., Ourednicek P., Lambert L. Deep convolutional neural network-based segmentation and classification of difficult to define metastatic spinal lesions in 3D CT data. Med Image Anal. 2018;49:76–88.
    1. Hay E.A., Parthasarathy R. Performance of convolutional neural networks for identification of bacteria in 3D microscopy datasets. PLoS Comput Biol. 2018;14
    1. Strodthoff N., Strodthoff C. Detecting and interpreting myocardial infarction using fully convolutional neural networks. Physiol Meas. 2018;40

Source: PubMed

3
Subscribe