Close

Loud and clear: AI is improving Assistant conversations

To get things done with the Google Assistant, it needs to understand you – it has to both recognize the words you’re saying, and also know what you mean. It should adapt to your way of talking, not require you to say exactly the right words in the right order.Understanding spoken language is difficult because…

To get things done with the Google Assistant, it needs to understand you – it has to both recognize the words you’re saying, and also know what you mean. It should adapt to your way of talking, not require you to say exactly the right words in the right order.

Understanding spoken language is difficult because it’s so contextual, and varies so much from person to person. And names can bring up other language hiccups — for instance, some names that are spelled the same are pronounced differently. It’s this kind of complexity that makes perfectly understanding the way we speak so difficult. This is something we’re working on with Assistant, and we have a few new improvements to share.

Teach Google to recognize unique names 

Names matter, and it’s frustrating when you’re trying to send a text or make a call and Google Assistant mispronounces or simply doesn’t recognize a contact. We want Assistant to accurately recognize and pronounce people’s names as often as possible, especially those that are less common.

Starting over the next few days, you can teach Google Assistant to enunciate and recognize names of your contacts the way you pronounce them. Assistant will listen to your pronunciation and remember it, without keeping a recording of your voice. This means Assistant will be able to better understand you when you say those names, and also be able to pronounce them correctly. The feature will be available in English and we hope to expand to more languages soon.

A good conversation is all about context

Assistant’s timers are a popular tool, and plenty of us set more than one of them at the same time. Maybe you’ve got a 10-minute timer for dinner going at the same time as another to remind the kids to start their homework in 20 minutes. You might fumble and stop mid sentence to correct how long the timer should be set for, or maybe you don’t use the exact same phrase to cancel it as you did to create it. Like in any conversation, context matters and Assistant needs to be flexible enough to understand what you’re referring to when you ask for help.

To help with these kinds of conversational complexities, we fully rebuilt Assistant’s NLU models so it can now more accurately understand context while also improving its “reference resolution” — meaning it knows exactly what you’re trying to do with a command.  This upgrade uses machine learning technology powered by state-of-the-art BERT, a technology we invented in 2018 and first brought to Search that makes it possible to process words in relation to all the other words in a sentence, rather than one-by-one in order. Because of these improvements, Assistant can now respond nearly 100 percent accurately to alarms and timer tasks. And over time, we’ll bring this capability to other use cases, so Assistant can learn to better understand you.

Source

Leave a Reply

Your email address will not be published. Required fields are marked *

0 Comments
scroll to top