Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings

Abstract Introduction Detecting voice disorders from voice recordings could allow for frequent, remote, and low-cost screening before costly clinical visits and a more invasive laryngoscopy examination. Our goals were to detect unilateral vocal fold paralysis (UVFP) from voice recordings using machine learning, to identify which acoustic variables were important for prediction to increase trust, and to determine model performance relative to clinician performance.Methods Patients with confirmed UVFP through endoscopic examination (N=77) and controls with normal voices matched for age and sex (N=77) were included. Voice samples were elicited by reading the Rainbow Passage and sustaining phonation of the vowel “a”. Four machine learning models of differing complexity were used. SHapley Additive exPlanations (SHAP) was used to identify important features.Results The highest median bootstrapped ROC AUC score was 0.87 and beat clinician’s performance (range: 0.74 – 0.81) based on the recordings. Recording durations were different between UVFP recordings and controls due to how that data was originally processed when storing, which we can show can classify both groups. And counterintuitively, many UVFP recordings had higher intensity than controls, when UVFP patients tend to have weaker voices, revealing a dataset-specific bias which we mitigate in an additional analysis.Conclusion We demonstrate that recording biases in audio duration and intensity created dataset-specific differences between patients and controls, which models used to improve classification. Furthermore, clinician’s ratings provide further evidence that patients were over-projecting their voices and being recorded at a higher amplitude signal than controls. Interestingly, after matching audio duration and removing variables associated with intensity in order to mitigate the biases, the models were able to achieve a similar high performance. We provide a set of recommendations to avoid bias when building and evaluating machine learning models for screening in laryngology..

Medienart:

Preprint

Erscheinungsjahr:

2024

Erschienen:

2024

Enthalten in:

bioRxiv.org - (2024) vom: 22. März Zur Gesamtaufnahme - year:2024

Sprache:

Englisch

Beteiligte Personen:

Low, Daniel M. [VerfasserIn]
Rao, Vishwanatha [VerfasserIn]
Randolph, Gregory [VerfasserIn]
Song, Phillip C. [VerfasserIn]
Ghosh, Satrajit S. [VerfasserIn]

Links:

Volltext [kostenfrei]

Themen:

570
Biology

doi:

10.1101/2020.11.23.20235945

funding:

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

XBI019409958