“We think this shows that the way you produce sound changes when you have COVID, even if you’re asymptomatic”.
How this model might work but why it might be very broken.
AI is not magic: AI simply builds up a rich representation of the world (a model) and uses this model to predict, categorise, and create. What is exciting about AI is the complex relationships which they can reflect.
AI cannot see all ends either.
AI can capture these relationships. So, if there is a relationship between the sound you produce when you force cough whilst healthy compared with force coughing when you have Covid-19, then AI can probably detect this difference. The question is: will there be a difference?
This is Biological plausibility. It is central to #AIinHealthcare because AI is not magic. AI cannot glance at tea leaves to predict how many people will attend my emergency department. Tea leaves (I’m 95% certain) have absolutely no relationship to these attendances.
However, if an AI can learn how combinations of factors conspire to affect a person’s decision to attend, then accurately it can predict.
So how does Covid-19 affect your cough?
SARS-Cov-2, the virus which causes Covid-19, enters into cells via a protein called ACE2.
Now, the ciliated epithelium (cells with hairs to move dust/dirt) of vocal cords is rich in ACE2. The squamous epithelium (other outer-layer cells around the cords) expresses less ACE2.
Image: Blausen.com staff (2014) “Medical gallery of Blausen Medical 2014”.
Drier vocal cords = change in forced cough.
But here is the problem: drier vocal cords = sore throat.
Forced cough should only be affected once symptoms develop.
Remember, cough samples were generated by people forcing themselves to cough.
Asymptomatic individuals were those without an involuntary cough or any other symptom. The AI is not listening passively to a pattern of coughing: that would be an entirely different model.
Other factors which may affect cough are muscle degradation (e.g coughing frequently, low energy from fighting the infection), vocal fatigue (again, from coughing or from shortness of breath), and emotional state. These features plausibly differentiate health from illness.
But hang on: the model performed *better* on asymptomatic individuals? This is where it falls apart for me. There are no factors which should affect forced-cough sounds which do not become better differentiated as symptoms develop. There are no mechanisms to explain this.
Bizarrely, the model performs worst on non-covid patients with Diarrhoea, misclassifying over 1 in 4 as having Covid. I can see no plausible explanation for how Diarrhoea could affect you cough in such a way as to make this model think you have Covid-19.
None of these mechanisms should exhibit any significant effects in asymptomatic individuals at all. So, what’s going on here? One explanation is bias using sentiment (emotion) in the model. These individuals knew their test results already, which could impact emotion.
But, reviewing the original paper: we see that sentiment contributed least to the model’s accuracy. The baseline model with just muscular degradation and the cough sound itself only performed at 60% accuracy. That’s not much better than a coin toss.
Vocal fatigue contributed 19% additional accuracy from baseline (getting 4 in 5 right). Biologically, this effect seems reasonable — but a miss-rate of 1 in 5 is difficult to tolerate in the real-world. Inability to differentiate between different illnesses is a weakness.
But the authors added in another biomarker, contributing a further 23% accuracy to the published 98.5%. Certainly an impressive feat: so what is this magic bullet? Well… it’s the ability to distinguish between people whose mother-tongue is Spanish vs English.
“We hypothesized that such models capable of learning features and acoustic variations on forced coughs trained to differentiate mother tongue could enhance COVID19 detection using transfer learning.”
Covid-19 affects people’s coughs differently based on their language?
Here’s what I think has actually happened (we would need the metadata from @MIT to know!): the data set is unbalanced with say, the majority of asymptomatic positive covid patients being Spanish speaking.
If the model learns to identify whether someone is Spanish or English, it is able to use this spurious correlation with asymptomatic +ve covidity to cheat and work out whether someone has covid. With only 100 asymptomatic positive patients in the data, this is important.
A similar situation occurred with a dataset I had for breath sounds for @Senti_Care. Hidden in the metadata was a ‘fact’ that only women get Asthma. This was simply a function of the dataset, of course, and has no resemblance to reality.
However, knowing a person’s sex gave the model tea leaves to decide whether a person didn’t have asthma. Auto-magically, it knows that, if you are a man, you don’t have Asthma. If this model made it out in the real world, where men have asthma too, it would fail hard.
This model therefore does have some potential: but it is not 98.5% and it is not in asymptomatic individuals. The authors need to check their data; broaden it to differentiate between influenza covid-19 and other illnesses, and test it in the real world.
Until then, the claim “We have proven COVID-19 can be discriminated with 98.5% accuracy using only a forced-cough and an AI biomarker focused approach.” is nothing more than hubris and hype.
“We find most remarkable that our model detected all of the COVID-19 positive asymptomatic patients, 100% of them.”
We find this remarkable too. Remarkable, and, unfortunately, implausible.
For more #AIinHealthCare pitfalls, check out our 10m read here: Creating real-world healthcare artificial intelligence
PS. If you can predict A&E attendances using tea leaves, then please comment here, make us a brew, and tell me what my future holds.