Abstract
With the COVID-19 pandemic, several research teams have reported successful advances in automated recognition of COVID-19 by voice. Resulting voice-based screening tools for COVID-19 could support large-scale testing efforts. While capabilities of machines on this task are progressing, we approach the so far unexplored aspect whether human raters can distinguish COVID-19 positive and negative tested speakers from voice samples, and compare their performance to a machine learning baseline. To account for the challenging symptom similarity between COVID-19 and other respiratory diseases, we use a carefully balanced dataset of voice samples, in which COVID-19 positive and negative tested speakers are matched by their symptoms alongside COVID-19 negative speakers without symptoms. Both human raters and the machine struggle to reliably identify COVID-19 positive speakers in our dataset. These results indicate that particular attention should be paid to the distribution of symptoms across all speakers of a dataset when assessing the capabilities of existing systems. The identification of acoustic aspects of COVID-19-related symptom manifestations might be the key for a reliable voice-based COVID-19 detection in the future by both trained human raters and machine learning models.
| Original language | English |
|---|---|
| Title of host publication | 22nd Annual Conference of the International Speech Communication Association, INTERSPEECH 2021 |
| Publisher | International Speech Communication Association |
| Pages | 701-705 |
| Number of pages | 5 |
| ISBN (Electronic) | 9781713836902 |
| DOIs | |
| State | Published - 2021 |
| Externally published | Yes |
| Event | 22nd Annual Conference of the International Speech Communication Association, INTERSPEECH 2021 - Brno, Czech Republic Duration: 30 Aug 2021 → 3 Sep 2021 |
Publication series
| Name | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH |
|---|---|
| Volume | 1 |
| ISSN (Print) | 2308-457X |
| ISSN (Electronic) | 2958-1796 |
Conference
| Conference | 22nd Annual Conference of the International Speech Communication Association, INTERSPEECH 2021 |
|---|---|
| Country/Territory | Czech Republic |
| City | Brno |
| Period | 30/08/21 → 3/09/21 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Keywords
- Auditory disease perception
- Automatic disease recognition
- Computational paralinguistics
- Covid-19
- Voice
Fingerprint
Dive into the research topics of 'Speaking Corona? Human and Machine Recognition of COVID-19 from Voice'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver