Please use this identifier to cite or link to this item: https://hdl.handle.net/10316/103499
Title: Automated Lung Sound Classification Using a Hybrid CNN-LSTM Network and Focal Loss Function
Authors: Petmezas, Georgios
Cheimariotis, Grigorios-Aris
Stefanopoulos, Leandros
Rocha, Bruno 
Paiva, Rui Pedro 
Katsaggelos, Aggelos K.
Maglaveras, Nicos
Keywords: lung sounds; crackles; wheezes; STFT; CNN; LSTM; focal loss; COPD; asthma
Issue Date: 6-Feb-2022
Publisher: MDPI
Project: EU-WELMO project (project number 210510516) 
Ph.D. scholarship SFRH/BD/135686/2018 
Ph.D. scholarship 2020.04927.BD 
metadata.degois.publication.title: Sensors
metadata.degois.publication.volume: 22
metadata.degois.publication.issue: 3
Abstract: Respiratory diseases constitute one of the leading causes of death worldwide and directly affect the patient's quality of life. Early diagnosis and patient monitoring, which conventionally include lung auscultation, are essential for the efficient management of respiratory diseases. Manual lung sound interpretation is a subjective and time-consuming process that requires high medical expertise. The capabilities that deep learning offers could be exploited in order that robust lung sound classification models can be designed. In this paper, we propose a novel hybrid neural model that implements the focal loss (FL) function to deal with training data imbalance. Features initially extracted from short-time Fourier transform (STFT) spectrograms via a convolutional neural network (CNN) are given as input to a long short-term memory (LSTM) network that memorizes the temporal dependencies between data and classifies four types of lung sounds, including normal, crackles, wheezes, and both crackles and wheezes. The model was trained and tested on the ICBHI 2017 Respiratory Sound Database and achieved state-of-the-art results using three different data splitting strategies-namely, sensitivity 47.37%, specificity 82.46%, score 64.92% and accuracy 73.69% for the official 60/40 split, sensitivity 52.78%, specificity 84.26%, score 68.52% and accuracy 76.39% using interpatient 10-fold cross validation, and sensitivity 60.29% and accuracy 74.57% using leave-one-out cross validation.
URI: https://hdl.handle.net/10316/103499
ISSN: 1424-8220
DOI: 10.3390/s22031232
Rights: openAccess
Appears in Collections:I&D CISUC - Artigos em Revistas Internacionais

Show full item record

SCOPUSTM   
Citations

80
checked on Nov 4, 2024

WEB OF SCIENCETM
Citations

48
checked on Nov 2, 2024

Page view(s)

121
checked on Nov 5, 2024

Download(s)

180
checked on Nov 5, 2024

Google ScholarTM

Check

Altmetric

Altmetric


This item is licensed under a Creative Commons License Creative Commons