For millions of folks around the world, speech impairment is a fact of life. That means it gives limited access to speech interfaces, such as digital assistants that depend on directly understanding one’s speech.
This poses a challenge for accessibility engineers developing AI-driven speech recognition and text-to-speech synthesis products, who have to accommodate a range of impairments for which limited data sets are available.
Fortunately, scientists at Google are investigating ways to minimize word substitution, deletion, and insertion errors in speech models as a part of Parrotron, an ongoing research initiative that aims to help those with atypical speech become better understood.
“Parrotron makes it easier for users with atypical speech to talk to and be understood by other people and by speech interfaces, with its end-to-end speech conversion approach more likely to reproduce the user’s intended speech.”
Thus, opening up the speech interface technology to all.