Respiratory sound classification for crackles, wheezes, and rhonchi in the clinical field using deep learning

Yoonjoo Kim, YunKyong Hyon, Sung Soo Jung, Sunju Lee, Geon Yoo, Chaeuk Chung, Taeyoung Ha, Yoonjoo Kim, YunKyong Hyon, Sung Soo Jung, Sunju Lee, Geon Yoo, Chaeuk Chung, Taeyoung Ha

Abstract

Auscultation has been essential part of the physical examination; this is non-invasive, real-time, and very informative. Detection of abnormal respiratory sounds with a stethoscope is important in diagnosing respiratory diseases and providing first aid. However, accurate interpretation of respiratory sounds requires clinician's considerable expertise, so trainees such as interns and residents sometimes misidentify respiratory sounds. To overcome such limitations, we tried to develop an automated classification of breath sounds. We utilized deep learning convolutional neural network (CNN) to categorize 1918 respiratory sounds (normal, crackles, wheezes, rhonchi) recorded in the clinical setting. We developed the predictive model for respiratory sound classification combining pretrained image feature extractor of series, respiratory sound, and CNN classifier. It detected abnormal sounds with an accuracy of 86.5% and the area under the ROC curve (AUC) of 0.93. It further classified abnormal lung sounds into crackles, wheezes, or rhonchi with an overall accuracy of 85.7% and a mean AUC of 0.92. On the other hand, as a result of respiratory sound classification by different groups showed varying degree in terms of accuracy; the overall accuracies were 60.3% for medical students, 53.4% for interns, 68.8% for residents, and 80.1% for fellows. Our deep learning-based classification would be able to complement the inaccuracies of clinicians' auscultation, and it may aid in the rapid diagnosis and appropriate treatment of respiratory diseases.

Conflict of interest statement

The authors declare no competing interests.

© 2021. The Author(s).

Figures

Figure 1
Figure 1
Scheme of the classification of respiratory sounds using deep learning. Lung sounds database contains normal sounds, crackles, wheezes, and rhonchi. Deep learning was used for two types of classification: The first step is the discriminating normal sounds from abnormal sounds. The second is to categorize abnormal sounds into crackles, wheezes, and rhonchi. (ER: Emergency room, ICU: intensive care unit).
Figure 2
Figure 2
ROC of the model for discrimination of abnormal lung sounds. Each plot illustrates the ROC of the algorithm on the independent testing set for abnormal lung sounds, with AUC of 0.93.
Figure 3
Figure 3
ROC of the model for classifying abnormal lung sounds into crackles, wheezes, and rhonchi. Each plot illustrates the ROC of the algorithm on the independent testing set for crackles, wheezes, and rhonchi with the mean AUC of 0.92.
Figure 4
Figure 4
Accuracy of auscultation analysis in real clinical practice. (A) Mean correction answer rates for the overall sounds, normal sounds, crackles, wheezes, and rhonchi. (B) Mean correction answer rates of students, interns, residents, and fellows for overall sounds. (C) Mean correction answer rates of students, interns, residents, and fellows for normal sounds (D) Mean correction answer rates of students, interns, residents, and fellows for crackles (E) Mean correction answer rates of students, interns, residents, and fellows for wheezes. (F) Mean correction answer rates of students, interns, residents, and fellows for rhonchi. *p < 0.05, **p < 0.05 ***p < 0.001 (Student's t-test).
Figure 5
Figure 5
Summary of deep learning assisted classification of respiratory sounds. Respiratory sounds were corrected from the patients with pulmonary diseases. The sounds were validated and classified by pulmonologists. The sounds were converted to Mel-spectrogram and features were extracted by VGG16 (transfer learning). Respiratory sounds were classified by CNN. Deep learning-based classification of respiratory sounds can be helpful for screening, monitoring, and diagnosis of pulmonary diseases.
Figure 6
Figure 6
Overview of our AI models.
Figure 7
Figure 7
Process to obtain spectrograms. (A) Given lung sound, dividing lung sound files with overlapping during 3 s (B) Obtaining three types of Mel-spectrograms with log-scale.
Figure 8
Figure 8
VGG16 architecture for our model. The input size for our model is 256 × 256. We freeze all layers in VGG16 without fully-connected layer, and with extracting features, we classified the respiratory sounds.
Figure 9
Figure 9
5-folds cross validation. Trying 5 iterations and getting final results with mean of all performances.

References

    1. Bohadana A, Izbicki G, Kraman SS. Fundamentals of lung auscultation. N. Engl. J. Med. 2014;370:744–751. doi: 10.1056/NEJMra1302901.
    1. Bloch H. The inventor of the stethoscope: René Laennec. J. Fam. Pract. 1993;37:191.
    1. Roguin A. Rene Theophile Hyacinthe Laennec (1781–1826): The man behind the stethoscope. Clin. Med. Res. 2006;4:230–235. doi: 10.3121/cmr.4.3.230.
    1. Swarup S, Makaryus AN. Digital stethoscope: Technology update. Med. Dev. (Auckland N. Z.) 2018;11:29–36. doi: 10.2147/mder.s135882.
    1. Leng S, et al. The electronic stethoscope. Biomed. Eng. Online. 2015;14:66. doi: 10.1186/s12938-015-0056-y.
    1. Arts L, Lim EHT, van de Ven PM, Heunks L, Tuinman PR. The diagnostic accuracy of lung auscultation in adult patients with acute pulmonary pathologies: A meta-analysis. Sci. Rep. 2020;10:7347. doi: 10.1038/s41598-020-64405-6.
    1. Sarkar M, Madabhavi I, Niranjan N, Dogra M. Auscultation of the respiratory system. Ann. Thoracic Med. 2015;10:158–168. doi: 10.4103/1817-1737.160831.
    1. Vyshedskiy A, et al. Mechanism of inspiratory and expiratory crackles. Chest. 2009;135:156–164. doi: 10.1378/chest.07-1562.
    1. Fukumitsu T, Obase Y, Ishimatsu Y. The acoustic characteristics of fine crackles predict honeycombing on high-resolution computed tomography. BMC Pulm. Med. 2019;19:153. doi: 10.1186/s12890-019-0916-5.
    1. Hafke-Dys H, Bręborowicz A, Kleka P. The accuracy of lung auscultation in the practice of physicians and medical students. PLoS ONE. 2019;14:e0220606. doi: 10.1371/journal.pone.0220606.
    1. Mangione S, Nieman LZ. Pulmonary auscultatory skills during training in internal medicine and family practice. Am. J. Respir. Crit. Care Med. 1999;159:1119–1124. doi: 10.1164/ajrccm.159.4.9806083.
    1. Melbye H, et al. Wheezes, crackles and rhonchi: Simplifying description of lung sounds increases the agreement on their classification: a study of 12 physicians' classification of lung sounds from video recordings. BMJ Open Respir. Res. 2016;3:e000136. doi: 10.1136/bmjresp-2016-000136.
    1. Andrès E, Gass R, Charloux A, Brandt C, Hentzler A. Respiratory sound analysis in the era of evidence-based medicine and the world of medicine 2.0. J. Med. Life. 2018;11:89–106.
    1. Ohshimo S, Sadamori T, Tanigawa K. Innovation in analysis of respiratory sounds. Ann. Intern. Med. 2016;164:638–639. doi: 10.7326/L15-0350.
    1. Altan G, Yayık A, Kutlu Y. Deep learning with ConvNet predicts imagery tasks through EEG. Neural Process. Lett. 2021;53:2917–2932. doi: 10.1007/s11063-021-10533-7.
    1. Tang YX, Tang YB, Peng Y. Automated abnormality classification of chest radiographs using deep convolutional neural networks. NPJ Dig. Med. 2020;3:70. doi: 10.1038/s41746-020-0273-z.
    1. Coucke PA. Laennec versus Forbes : Tied for the score ! How technology helps us interpret auscultation. Rev. Med. Liege. 2019;74:543–551.
    1. Grzywalski T, et al. Practical implementation of artificial intelligence algorithms in pulmonary auscultation examination. Eur. J. Pediatr. 2019;178:883–890. doi: 10.1007/s00431-019-03363-2.
    1. Palaniappan R, Sundaraj K, Sundaraj S. Artificial intelligence techniques used in respiratory sound analysis: A systematic review. Biomedizinische Technik. Biomed. Eng. 2014;59:7–18. doi: 10.1515/bmt-2013-0074.
    1. Ono H, et al. Evaluation of the usefulness of spectral analysis of inspiratory lung sounds recorded with phonopneumography in patients with interstitial pneumonia. J. Nippon Med. School Nippon Ika Daigaku Zasshi. 2009;76:67–75. doi: 10.1272/jnms.76.67.
    1. Thompson WR, Reinisch AJ, Unterberger MJ, Schriefl AJ. Artificial intelligence-assisted auscultation of heart murmurs: Validation by virtual clinical trial. Pediatr. Cardiol. 2019;40:623–629. doi: 10.1007/s00246-018-2036-z.
    1. Murphy RL, et al. Automated lung sound analysis in patients with pneumonia. Respir. Care. 2004;49:1490–1497.
    1. Kevat A, Kalirajah A, Roseby R. Artificial intelligence accuracy in detecting pathological breath sounds in children using digital stethoscopes. Respir. Res. 2020;21:253. doi: 10.1186/s12931-020-01523-9.
    1. Serbes, G., Sakar, C. O., Kahya, Y. P. & Aydin, N. Feature extraction using time-frequency/scale analysis and ensemble of feature sets for crackle detection. In Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference2011, 3314–3317. 10.1109/IEMBS.2011.6090899 (2011).
    1. Demir F, Sengur A, Bajaj V. Convolutional neural networks based efficient approach for classification of lung diseases. Health Inform. Sci. Syst. 2020;8:4. doi: 10.1007/s13755-019-0091-3.
    1. Chamberlain, D., Kodgule, R., Ganelin, D., Miglani, V. & Fletcher, R. R. Application of semi-supervised deep learning to lung sound analysis. In Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference2016, 804–807. 10.1109/EMBC.2016.7590823 (2016).
    1. Guler I, Polat H, Ergun U. Combining neural network and genetic algorithm for prediction of lung sounds. J. Med. Syst. 2005;29:217–231. doi: 10.1007/s10916-005-5182-9.
    1. Altan G, Kutlu Y, Allahverdi N. Deep learning on computerized analysis of chronic obstructive pulmonary disease. IEEE J. Biomed. Health Inform. 2019 doi: 10.1109/JBHI.2019.2931395.
    1. Altan G, Kutlu Y, Pekmezci AÖ, Nural S. Deep learning with 3D-second order difference plot on respiratory sounds. Biomed. Signal Process. Control. 2018;45:58–69. doi: 10.1016/j.bspc.2018.05.014.
    1. Altan G, Kutlu Y, Gökçen A. Chronic obstructive pulmonary disease severity analysis using deep learning on multi-channel lung sounds. Turk. J. Electr. Eng. Comput. Sci. 2020;28:2979–2996. doi: 10.3906/elk-2004-68.
    1. Altan G, Kutlu Y, Garbi Y, Pekmezci AÖ, Nural S. Multimedia respiratory database (RespiratoryDatabase@ TR): Auscultation sounds and chest X-rays. Nat. Eng. Sci. 2017;2:59–72.
    1. Aras S, Öztürk M, Gangal A. Automatic detection of the respiratory cycle from recorded, single-channel sounds from lungs. Turk. J. Electr. Eng. Comput. Sci. 2018;26:11–22. doi: 10.3906/elk-1705-16.
    1. Zheng L, et al. Artificial intelligence in digital cariology: A new tool for the diagnosis of deep caries and pulpitis using convolutional neural networks. Ann. Transl. Med. 2021;9:763. doi: 10.21037/atm-21-119.
    1. Rezaeijo SM, Ghorvei M, Mofid B. Predicting breast cancer response to neoadjuvant chemotherapy using ensemble deep transfer learning based on CT images. J. Xray Sci. Technol. 2021 doi: 10.3233/xst-210910.
    1. Arora V, Ng EY, Leekha RS, Darshan M, Singh A. Transfer learning-based approach for detecting COVID-19 ailment in lung CT scan. Comput. Biol. Med. 2021;135:104575. doi: 10.1016/j.compbiomed.2021.104575.
    1. Jin W, Dong S, Dong C, Ye X. Hybrid ensemble model for differential diagnosis between COVID-19 and common viral pneumonia by chest X-ray radiograph. Comput. Biol. Med. 2021;131:104252. doi: 10.1016/j.compbiomed.2021.104252.
    1. Wang Q, et al. Realistic lung nodule synthesis with multi-target co-guided adversarial mechanism. IEEE Trans. Med. Imag. 2021 doi: 10.1109/tmi.2021.3077089.
    1. Pu J, Sechrist J, Meng X, Leader JK, Sciurba FC. A pilot study: Quantify lung volume and emphysema extent directly from two-dimensional scout images. Med. Phys. 2021 doi: 10.1002/mp.15019.
    1. Gharehbaghi A, Linden M. A deep machine learning method for classifying cyclic time series of biological signals using time-growing neural network. IEEE Trans. Neural Netw. Learn. Syst. 2018;29:4102–4115. doi: 10.1109/TNNLS.2017.2754294.
    1. Raghu, M., Chiyuan, Z., Jon, K. & Samy, B. Transfusion: Understanding Transfer learning for medical imaging. In 33rd Conference on Neural Information Processing Systems (2019).
    1. Park S, Kim J, Kim D. A study on classification performance analysis of convolutional neural network using ensemble learning algorithm. J. Korea Multimedia Soc. 2019;22:665–675.
    1. Epler GR, Carrington CB, Gaensler EA. Crackles (rales) in the interstitial pulmonary diseases. Chest. 1978;73:333–339. doi: 10.1378/chest.73.3.333.
    1. Horimasu Y, et al. A machine-learning based approach to quantify fine crackles in the diagnosis of interstitial pneumonia: A proof-of-concept study. Medicine. 2021;100:e24738. doi: 10.1097/md.0000000000024738.
    1. Naves R, Barbosa BH, Ferreira DD. Classification of lung sounds using higher-order statistics: A divide-and-conquer approach. Comput. Methods Programs Biomed. 2016;129:12–20. doi: 10.1016/j.cmpb.2016.02.013.
    1. Aykanat M, Kılıç Ö, Kurt B, Saryal S. Classification of lung sounds using convolutional neural networks. J. Image Video Proc. 2017 doi: 10.1186/s13640-017-0213-2.
    1. McFee, B. librosa. librosa 0.8.0. 10.5281/zenodo.3955228 (2020).
    1. Bardou D, Zhang K, Ahmad SM. Lung sounds classification using convolutional neural networks. Artif. Intell. Med. 2018;88:58–59. doi: 10.1016/j.artmed.2018.04.008(2018).
    1. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv. Neural Inform. Process. Syst. 2012;25:1097–1105.
    1. Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. CoRR abs1409, 1556 (2014).

Source: PubMed

3
Sottoscrivi