Multi-type outer product-based fusion of respiratory sounds for detecting COVID-19

Mostra el registre complet Registre parcial de l'ítem

  • dc.contributor.author Mallol Ragolta, Adrià
  • dc.contributor.author Cuesta, Helena
  • dc.contributor.author Gómez Gutiérrez, Emilia, 1975-
  • dc.contributor.author Schuller, Björn
  • dc.date.accessioned 2023-03-01T13:48:11Z
  • dc.date.available 2023-03-01T13:48:11Z
  • dc.date.issued 2022
  • dc.description Comunicació presentada a Interspeech 2022, celebrat del 18 al 22 de setembre de 2022 a Inchon, Corea del Sud.
  • dc.description.abstract This work presents an outer product-based approach to fuse the embedded representations learnt from the spectrograms of cough, breath, and speech samples for the automatic detection of COVID-19. To extract deep learnt representations from the spectrograms, we compare the performance of specific Convolutional Neural Networks (CNNs) trained from scratch and ResNet18- based CNNs fine-tuned for the task at hand. Furthermore, we investigate whether the patients’ sex and the use of contextual attention mechanisms are beneficial. Our experiments use the dataset released as part of the Second Diagnosing COVID-19 using Acoustics (DiCOVA) Challenge. The results suggest the suitability of fusing breath and speech information to detect COVID-19. An Area Under the Curve (AUC) of 84.06 % is obtained on the test partition when using specific CNNs trained from scratch with contextual attention mechanisms. When using ResNet18-based CNNs for feature extraction, the baseline model scores the highest performance with an AUC of 84.26 %.
  • dc.description.sponsorship This project has received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement No. 826506 (sustAGE), and from the Spanish Ministry of Science and Innovation under the Musical AI project (PID2019-111403GB-I00).
  • dc.format.mimetype application/pdf
  • dc.identifier.citation Mallol-Ragolta A, Cuesta H, Gomez E, Schuller B. Multi-type outer product-based fusion of respiratory sounds for detecting COVID-19. In: Ko H, Hansen JHL. Interspeech 2022; 2022 Sep 18-22; Incheon, Korea. [Baixas]: International Speech Communication Association; 2022. p. 2163-7. DOI: 10.21437/Interspeech.2022-10291
  • dc.identifier.doi http://dx.doi.org/10.21437/Interspeech.2022-10291
  • dc.identifier.uri http://hdl.handle.net/10230/55986
  • dc.language.iso eng
  • dc.publisher International Speech Communication Association (ISCA)
  • dc.relation.ispartof Ko H, Hansen JHL. Interspeech 2022; 2022 Sep 18-22; Incheon, Korea. [Baixas]: International Speech Communication Association; 2022. p. 2163-7.
  • dc.relation.projectID info:eu-repo/grantAgreement/EC/H2020/826506
  • dc.relation.projectID info:eu-repo/grantAgreement/ES/2PE/PID2019-111403GB-I00
  • dc.rights © 2022 ISCA
  • dc.rights.accessRights info:eu-repo/semantics/openAccess
  • dc.subject.keyword COVID-19 Detection
  • dc.subject.keyword Respiratory Diagnosis
  • dc.subject.keyword Transfer Learning
  • dc.subject.keyword Information Fusion
  • dc.subject.keyword Healthcare
  • dc.title Multi-type outer product-based fusion of respiratory sounds for detecting COVID-19
  • dc.type info:eu-repo/semantics/conferenceObject
  • dc.type.version info:eu-repo/semantics/publishedVersion