Home > Research > Publications & Outputs > Audio-based AI classifiers show no evidence of ...

Links

Text available via DOI:

View graph of relations

Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers. / Coppock, H.; Nicholson, G.; Kiskin, I. et al.
In: Nature Machine Intelligence, Vol. 6, 07.02.2024, p. 229-242.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

Coppock, H, Nicholson, G, Kiskin, I, Koutra, V, Baker, K, Budd, J, Payne, R, Karoune, E, Hurley, D, Titcomb, A, Egglestone, S, Tendero Cañadas, A, Butler, L, Jersakova, R, Mellor, J, Patel, S, Thornley, T, Diggle, P, Richardson, S, Packham, J, Schuller, BW, Pigoli, D, Gilmour, S, Roberts, S & Holmes, C 2024, 'Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers', Nature Machine Intelligence, vol. 6, pp. 229-242. https://doi.org/10.1038/s42256-023-00773-8

APA

Coppock, H., Nicholson, G., Kiskin, I., Koutra, V., Baker, K., Budd, J., Payne, R., Karoune, E., Hurley, D., Titcomb, A., Egglestone, S., Tendero Cañadas, A., Butler, L., Jersakova, R., Mellor, J., Patel, S., Thornley, T., Diggle, P., Richardson, S., ... Holmes, C. (2024). Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers. Nature Machine Intelligence, 6, 229-242. https://doi.org/10.1038/s42256-023-00773-8

Vancouver

Coppock H, Nicholson G, Kiskin I, Koutra V, Baker K, Budd J et al. Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers. Nature Machine Intelligence. 2024 Feb 7;6:229-242. doi: 10.1038/s42256-023-00773-8

Author

Coppock, H. ; Nicholson, G. ; Kiskin, I. et al. / Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers. In: Nature Machine Intelligence. 2024 ; Vol. 6. pp. 229-242.

Bibtex

@article{7eaba98ba349489cacd560d1b4768ed1,
title = "Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers",
abstract = "Recent work has reported that respiratory audio-trained AI classifiers can accurately predict SARS-CoV-2 infection status. However, it has not yet been determined whether such model performance is driven by latent audio biomarkers with true causal links to SARS-CoV-2 infection or by confounding effects, such as recruitment bias, present in observational studies. Here we undertake a large-scale study of audio-based AI classifiers as part of the UK government{\textquoteright}s pandemic response. We collect a dataset of audio recordings from 67,842 individuals, with linked metadata, of whom 23,514 had positive polymerase chain reaction tests for SARS-CoV-2. In an unadjusted analysis, similar to that in previous works, AI classifiers predict SARS-CoV-2 infection status with high accuracy (ROC–AUC = 0.846 [0.838–0.854]). However, after matching on measured confounders, such as self-reported symptoms, performance is much weaker (ROC–AUC = 0.619 [0.594–0.644]). Upon quantifying the utility of audio-based classifiers in practical settings, we find them to be outperformed by predictions on the basis of user-reported symptoms. We make best-practice recommendations for handling recruitment bias, and for assessing audio-based classifiers by their utility in relevant practical settings. Our work provides insights into the value of AI audio analysis and the importance of study design and treatment of confounders in AI-enabled diagnostics.",
author = "H. Coppock and G. Nicholson and I. Kiskin and V. Koutra and K. Baker and J. Budd and R. Payne and E. Karoune and D. Hurley and A. Titcomb and S. Egglestone and {Tendero Ca{\~n}adas}, A. and L. Butler and R. Jersakova and J. Mellor and S. Patel and T. Thornley and P. Diggle and S. Richardson and J. Packham and B.W. Schuller and D. Pigoli and S. Gilmour and S. Roberts and C. Holmes",
year = "2024",
month = feb,
day = "7",
doi = "10.1038/s42256-023-00773-8",
language = "English",
volume = "6",
pages = "229--242",
journal = "Nature Machine Intelligence",
issn = "2522-5839",
publisher = "Springer Science and Business Media LLC",

}

RIS

TY - JOUR

T1 - Audio-based AI classifiers show no evidence of improved COVID-19 screening over simple symptoms checkers

AU - Coppock, H.

AU - Nicholson, G.

AU - Kiskin, I.

AU - Koutra, V.

AU - Baker, K.

AU - Budd, J.

AU - Payne, R.

AU - Karoune, E.

AU - Hurley, D.

AU - Titcomb, A.

AU - Egglestone, S.

AU - Tendero Cañadas, A.

AU - Butler, L.

AU - Jersakova, R.

AU - Mellor, J.

AU - Patel, S.

AU - Thornley, T.

AU - Diggle, P.

AU - Richardson, S.

AU - Packham, J.

AU - Schuller, B.W.

AU - Pigoli, D.

AU - Gilmour, S.

AU - Roberts, S.

AU - Holmes, C.

PY - 2024/2/7

Y1 - 2024/2/7

N2 - Recent work has reported that respiratory audio-trained AI classifiers can accurately predict SARS-CoV-2 infection status. However, it has not yet been determined whether such model performance is driven by latent audio biomarkers with true causal links to SARS-CoV-2 infection or by confounding effects, such as recruitment bias, present in observational studies. Here we undertake a large-scale study of audio-based AI classifiers as part of the UK government’s pandemic response. We collect a dataset of audio recordings from 67,842 individuals, with linked metadata, of whom 23,514 had positive polymerase chain reaction tests for SARS-CoV-2. In an unadjusted analysis, similar to that in previous works, AI classifiers predict SARS-CoV-2 infection status with high accuracy (ROC–AUC = 0.846 [0.838–0.854]). However, after matching on measured confounders, such as self-reported symptoms, performance is much weaker (ROC–AUC = 0.619 [0.594–0.644]). Upon quantifying the utility of audio-based classifiers in practical settings, we find them to be outperformed by predictions on the basis of user-reported symptoms. We make best-practice recommendations for handling recruitment bias, and for assessing audio-based classifiers by their utility in relevant practical settings. Our work provides insights into the value of AI audio analysis and the importance of study design and treatment of confounders in AI-enabled diagnostics.

AB - Recent work has reported that respiratory audio-trained AI classifiers can accurately predict SARS-CoV-2 infection status. However, it has not yet been determined whether such model performance is driven by latent audio biomarkers with true causal links to SARS-CoV-2 infection or by confounding effects, such as recruitment bias, present in observational studies. Here we undertake a large-scale study of audio-based AI classifiers as part of the UK government’s pandemic response. We collect a dataset of audio recordings from 67,842 individuals, with linked metadata, of whom 23,514 had positive polymerase chain reaction tests for SARS-CoV-2. In an unadjusted analysis, similar to that in previous works, AI classifiers predict SARS-CoV-2 infection status with high accuracy (ROC–AUC = 0.846 [0.838–0.854]). However, after matching on measured confounders, such as self-reported symptoms, performance is much weaker (ROC–AUC = 0.619 [0.594–0.644]). Upon quantifying the utility of audio-based classifiers in practical settings, we find them to be outperformed by predictions on the basis of user-reported symptoms. We make best-practice recommendations for handling recruitment bias, and for assessing audio-based classifiers by their utility in relevant practical settings. Our work provides insights into the value of AI audio analysis and the importance of study design and treatment of confounders in AI-enabled diagnostics.

U2 - 10.1038/s42256-023-00773-8

DO - 10.1038/s42256-023-00773-8

M3 - Journal article

VL - 6

SP - 229

EP - 242

JO - Nature Machine Intelligence

JF - Nature Machine Intelligence

SN - 2522-5839

ER -