The number of ways to pronounce “water” may be as ubiquitous as the substance itself. In 2003, a team of Harvard University researchers concluded a survey and mapped out the idiosyncrasies in the pronunciation of common English words across the globe. (How do you say “pecan?”)
But just because the same word may sound funny—perhaps even unintelligible—in different parts of the world doesn’t mean the speaker is wrong. Yet some of the voice recognition systems today were neither designed nor trained to pick up different dialects, and that can be a problem when they are used to assess oral fluency, says Patricia Scanlon.
“I can drive three counties away, and probably won’t understand some of the words people say,” quips Scanlon, the founder and CEO of SoapBox Labs, who is based in Dublin, Ireland. “But that doesn’t mean they’re wrong.”
Scanlon has plenty to say about speech. She is an engineer who holds a doctorate in speech-recognition technology, and has spent the past two decades researching this field at Columbia University, IBM and Bell Laboratories.
In 2013, after watching her three-year-old daughter work through phonics exercises on an iPad, it struck Scanlon that the apps at the time often did poorly measuring oral fluency. Some offered multiple-choice questions, for instance, that simply asked kids to pick the correct pronunciation—hardly a useful exercise.