Please use this identifier to cite or link to this item: http://repository.aaup.edu/jspui/handle/123456789/1708
Full metadata record
DC FieldValueLanguage
dc.contributor.authorZayed, Yara$AAUP$Palestinian-
dc.contributor.authorHasasneh, Ahmad$AAUP$Palestinian-
dc.contributor.authorTadj, Chakib$Other$Other-
dc.date.accessioned2023-10-05T07:51:42Z-
dc.date.available2023-10-05T07:51:42Z-
dc.date.issued2023-06-19-
dc.identifier.citationZayed, Y.; Hasasneh, A.; Tadj, C. Infant Cry Signal Diagnostic System Using Deep Learning and Fused Features. Diagnostics 2023, 13, 2107. https://doi.org/10.3390/diagnostics13122107en_US
dc.identifier.urihttp://repository.aaup.edu/jspui/handle/123456789/1708-
dc.description.abstractEarly diagnosis of medical conditions in infants is crucial for ensuring timely and effective treatment. However, infants are unable to verbalize their symptoms, making it difficult for healthcare professionals to accurately diagnose their conditions. Crying is often the only way for infants to communicate their needs and discomfort. In this paper, we propose a medical diagnostic system for interpreting infants’ cry audio signals (CAS) using a combination of different audio domain features and deep learning (DL) algorithms. The proposed system utilizes a dataset of labeled audio signals from infants with specific pathologies. The dataset includes two infant pathologies with high mortality rates, neonatal respiratory distress syndrome (RDS), sepsis, and crying. The system employed the harmonic ratio (HR) as a prosodic feature, the Gammatone frequency cepstral coefficients (GFCCs) as a cepstral feature, and image-based features through the spectrogram which are extracted using a convolution neural network (CNN) pretrained model and fused with the other features to benefit multiple domains in improving the classification rate and the accuracy of the model. The different combination of the fused features is then fed into multiple machine learning algorithms including random forest (RF), support vector machine (SVM), and deep neural network (DNN) models. The evaluation of the system using the accuracy, precision, recall, F1-score, confusion matrix, and receiver operating characteristic (ROC) curve, showed promising results for the early diagnosis of medical conditions in infants based on the crying signals only, where the system achieved the highest accuracy of 97.50% using the combination of the spectrogram, HR, and GFCC through the deep learning process. The finding demonstrated the importance of fusing different audio features, especially the spectrogram, through the learning process rather than a simple concatenation and the use of deep learning algorithms in extracting sparsely represented features that can be used later on in the classification problem, which improves the separation between different infants’ pathologies. The results outperformed the published benchmark paper by improving the classification problem to be multiclassification (RDS, sepsis, and healthy), investigating a new type of feature, which is the spectrogram, using a new feature fusion technique, which is fusion, through the learning process using the deep learning model.en_US
dc.language.isoenen_US
dc.publisherDiagnostics - MDPIen_US
dc.subjectinfant’s crying diagnosisen_US
dc.subjectaudio domains featuresen_US
dc.subjectHRen_US
dc.subjectGFCCen_US
dc.subjectmachine learningen_US
dc.subjectdeep learningen_US
dc.subjectspectrogramen_US
dc.titleInfant Cry Signal Diagnostic System Using Deep Learning and Fused Featuresen_US
dc.typeArticleen_US
Appears in Collections:Faculty & Staff Scientific Research publications

Files in This Item:
File Description SizeFormat 
diagnostics-13-02107 (1).pdf6.53 MBAdobe PDFView/Open
Show simple item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Admin Tools