15 February 2017 – In between flights I like to catch-up with New Scientist magazine … I travel with a stack jammed into my carry-on … and one article especially caught my eye.
A U.S. start-up called Canary Speech is developing deep-learning algorithms to detect if people have neurological conditions like Parkinson’s or Alzheimer’s disease just by listening to the sound of their voice. And it’s found a controversial source of audio data to train its algorithms on: phone calls to a health insurer.
The health insurer – which Canary Speech would not name but says is “a very large American healthcare and insurance provider” – has provided the company with hundreds of millions of phone calls that have been collected over the past 15 years and are labelled with information about the speaker’s medical history and demographic background.
Using this data, the company says its algorithms could pick up on vocal cues that distinguish someone with a particular condition from someone without that condition. For modelling purposes, they want to be able to see an individual over a period of years. The founders say the company hasn’t yet received all of the audio data, but could have an algorithm that aims to detect vocal indicators of Alzheimer’s disease ready within two months. It also aims to look for vocal markers for depression, stress and dyslexia.
All about the data
While you may be surprised to learn that calls to a health insurer could be used in this way. Canary Speech is working with has “express permissions” in place to allow it. (They have spoken to several UK companies too, but that stricter UK data protection laws may prevent a similar project there.)
How the technology is ultimately used will be down to Canary Speech’s customers. The insurer that is providing the call data also manages clinics in the US and discussions have so far related to using the technology in a clinical setting, for example to help with diagnosing these conditions. When asked if the technology could potentially be used to screen callers or influence insurance premiums, the article notes he responded by saying that such an application “may be regulated”. The article also quotes Caitlin Donovan at the US National Patient Advocate Foundation who says “this is the type of thing where we’d want to make sure patient privacy is protected. I would be worried that this algorithm could be used to either track or diagnose a condition that the patient may not be aware they have.
NOTE: the U.S. Affordable Care Act (ACA) prevents insurers from denying benefits or raising costs because of a pre-existing condition, but that could change under the new administration of Donald Trump. Before the ACA took effect, people with Parkinson’s and Alzheimer’s were often denied coverage by insurers.
Early diagnosis
We have known about vocal indicators of neurological conditions for a long time, says Sandy Schneider at Saint Mary’s College in Notre Dame, Indiana. But only recently have people started to explore how machine-learning could aid diagnoses:
Right now, it takes somebody maybe two or three years to be definitively diagnosed with a neurodegenerative disease. Algorithms may one day be able to detect symptoms in the voice before clinical diagnosis, meaning treatment of symptoms could start sooner.
I did some side-reading and found out the Canary Speech team is coming at the problem from a big data perspective, rather than a medical one. They have a large collection of audio files of people with and without diseases, and they just have to find in those files what are the differences in the audio, in the waves. It doesn’t require medical intuition to do that, it requires signal processing and machine learning.
The New Scientist piece also quotes Max Little at Aston University in Birmingham, UK, who explores the use of voice-processing algorithms to predict the severity of Parkinson’s, and says he isn’t convinced by the big data approach: “I’m deeply skeptical of efforts that naively believe you can simply accumulate more data and that will do the job”. Little uses data collected from relatively small numbers of patients who are recorded while they read or repeat specific sounds or phrases. Machine-learning algorithms then search for vocal indicators that are associated with Parkinson’s, such as hypophonia, a softness of speech resulting from lack of coordination over the vocal muscles.
Canary Speech, however, is feeding its machine-learning algorithms with vast amounts of conversational speech, in the hope they will learn to recognize subtle differences in the voices of people with different conditions. Says Little:
People have been able to show that there is enormous promise in this approach. But to really scale it up to be large enough that you could basically throw away all of the fundamental science and just go with some machine learning algorithm – I’m not so sure about that.