Google opens new communication channels for people with disabilities

Speech recognition has evolved rapidly in recent years. However, it remains difficult for people with speech impairments to communicate. wants to change this with Project Euphonia.

Who of you already uses language assistants like the Google Assistant? Whether you use them regularly or only now and then, you will have certainly been annoyed more than once about the fact that it simply did not understand what one wanted from it. Now imagine you were trying to communicate to a bot with a speaking impairment.

For people who cannot communicate like many others due to stroke, multiple sclerosis, ALS, Parkinson's disease or any other neurological or motor disorder, Google can help. After all, , especially verbal , is a significant part of our lives. To address this problem from a tech angle, Google has created Project Euphonia as part of its Social AI team.

Project Euphonia

The aim of Project Euphonia is to use AI to improve speech recognition and the recognition of facial expressions and gestures. For example, they have teamed up with the ALS Therapy Development Institute (ALS TDI) and the ALS Residence Initiative (ALSRI) to record numerous people with ALS and use them as a basis for improving speech recognition.

Euphonia Waveform example sentences 01

During recognition, Google's software converts the recorded speech samples into a spectrogram or a visual representation of the sounds. The whole thing is then converted into the usual spectrogram to “train” the system, which is then supposed to recognize this less common type of speech better.

In the following video, you can see Dimitri Kanevsky, a linguist who learned English from this base after becoming deaf as a child in Russia. He already uses a live translation program that was trained with the help of Project Euphonia and helps him to communicate.


As seen in the video, it is also important to be able to recognize facial expressions, gestures and even sounds such as humming or clicking with the tongue in order, for example, to continue to provide an opportunity for communication in advanced MS people. Steve Saling, who suffers from ALS, for example, uses noise detection to control various smart home devices.

You might also like

Comments are closed.

This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. AcceptRead More