It seems we’ll be seeing these improvements announced steadily over the coming year, as we surpass human WER parity and then start to get better at particular domains. Layering on top of this noise rejection, far field performance, and augmentation through training, we’re within two years of being understood better by a machine than by each other.
When that happens, we may start to rely on tools to analyze voice for us to give us feedback on things like truthfulness, emotion, or other measures to enhance our understanding of what people are telling us.
We just sent you an email. Please click the link in the email to confirm your subscription!