18382886. AUTOMATICALLY DETERMINING LANGUAGE FOR SPEECH RECOGNITION OF SPOKEN UTTERANCE RECEIVED VIA AN AUTOMATED ASSISTANT INTERFACE simplified abstract (GOOGLE LLC)

From WikiPatents
Jump to navigation Jump to search

AUTOMATICALLY DETERMINING LANGUAGE FOR SPEECH RECOGNITION OF SPOKEN UTTERANCE RECEIVED VIA AN AUTOMATED ASSISTANT INTERFACE

Organization Name

GOOGLE LLC

Inventor(s)

Pu-sen Chao of Los Altos CA (US)

Diego Melendo Casado of Mountain View CA (US)

Ignacio Lopez Moreno of New York NY (US)

AUTOMATICALLY DETERMINING LANGUAGE FOR SPEECH RECOGNITION OF SPOKEN UTTERANCE RECEIVED VIA AN AUTOMATED ASSISTANT INTERFACE - A simplified explanation of the abstract

This abstract first appeared for US patent application 18382886 titled 'AUTOMATICALLY DETERMINING LANGUAGE FOR SPEECH RECOGNITION OF SPOKEN UTTERANCE RECEIVED VIA AN AUTOMATED ASSISTANT INTERFACE

Simplified Explanation

The patent application describes a method for determining the language for speech recognition of a spoken utterance received via an automated assistant interface.

  • User profiles are created based on audio data capturing spoken utterances.
  • Languages and probabilities are assigned to user profiles.
  • The system selects languages from the user profile for speech recognition.
  • Criteria are used to select the appropriate speech recognition for generating responsive content.

---

      1. Potential Applications
  • Multilingual interaction with automated assistants
  • Seamless language detection for speech recognition
      1. Problems Solved
  • Eliminates the need for users to explicitly designate a language for each interaction
  • Enables automated assistants to understand and respond in multiple languages without user intervention
      1. Benefits
  • Improved user experience with automated assistants
  • Efficient and accurate language detection for speech recognition
  • Enhanced multilingual support for a wide range of users


Original Abstract Submitted

Determining a language for speech recognition of a spoken utterance received via an automated assistant interface for interacting with an automated assistant. Implementations can enable multilingual interaction with the automated assistant, without necessitating a user explicitly designate a language to be utilized for each interaction. Implementations determine a user profile that corresponds to audio data that captures a spoken utterance, and utilize language(s), and optionally corresponding probabilities, assigned to the user profile in determining a language for speech recognition of the spoken utterance. Some implementations select only a subset of languages, assigned to the user profile, to utilize in speech recognition of a given spoken utterance of the user. Some implementations perform speech recognition in each of multiple languages assigned to the user profile, and utilize criteria to select only one of the speech recognitions as appropriate for generating and providing content that is responsive to the spoken utterance.