US Patent Application 17726465. INTELLIGENT DISPLAY OF AUDITORY WORLD EXPERIENCES simplified abstract

From WikiPatents
Jump to navigation Jump to search

INTELLIGENT DISPLAY OF AUDITORY WORLD EXPERIENCES

Organization Name

Microsoft Technology Licensing, LLC


Inventor(s)

Venkata Naga Vijaya Swetha Machanavajhala of Redmond WA (US)


Ryan Graham Williams of Duvall WA (US)


Sanghee Oh of Los Angeles CA (US)


Ikuyo Tsunoda of Seattle WA (US)


William D. Lewis of Seattle WA (US)


Jian Wu of Bellevue WA (US)


Daniel Charles Tompkins of Bellevue WA (US)


INTELLIGENT DISPLAY OF AUDITORY WORLD EXPERIENCES - A simplified explanation of the abstract

  • This abstract for appeared for US patent application number 17726465 Titled 'INTELLIGENT DISPLAY OF AUDITORY WORLD EXPERIENCES'

Simplified Explanation

The techniques described in this abstract involve using specialized artificial intelligence models to display visual representations of auditory experiences. These models can analyze various aspects of speech, such as volume and tone, to identify specific characteristics. They can also recognize keywords in the speech to distinguish different parts of a transcript. Additionally, the models can analyze non-speech audio sounds to identify non-speech events. The system then combines these analyses with user interface attributes to provide visual indicators for the different aspects of the auditory signals.


Original Abstract Submitted

The techniques disclosed herein provide intelligent display of auditory world experiences. Specialized Al models are configured to display integrated visualizations for different aspects of the auditory signals that may be communicated during an event, such as a meeting, chat session, etc. For instance, a system can use a sentiment recognition model to identify specific characteristics of a speech input, such as volume or tone, provided by a participant. The system can also use a speech recognition model to identify keywords that can be used to distinguish portions of a transcript that are displayed. The system can also utilize an audio recognition model that is configured to analyze non-speech audio sounds for the purposes of identifying non-speech events. The system can then integrate the user interface attributes, distinguished portions of the transcript, and visual indicators describing the non-speech events.