US Patent Application 17827588. AUTOMATION OF VISUAL INDICATORS FOR DISTINGUISHING ACTIVE SPEAKERS OF USERS DISPLAYED AS THREE-DIMENSIONAL REPRESENTATIONS simplified abstract

From WikiPatents
Jump to navigation Jump to search

AUTOMATION OF VISUAL INDICATORS FOR DISTINGUISHING ACTIVE SPEAKERS OF USERS DISPLAYED AS THREE-DIMENSIONAL REPRESENTATIONS

Organization Name

Microsoft Technology Licensing, LLC==Inventor(s)==

[[Category:Jason Thomas Faulkner of Seattle WA (US)]]

AUTOMATION OF VISUAL INDICATORS FOR DISTINGUISHING ACTIVE SPEAKERS OF USERS DISPLAYED AS THREE-DIMENSIONAL REPRESENTATIONS - A simplified explanation of the abstract

This abstract first appeared for US patent application 17827588 titled 'AUTOMATION OF VISUAL INDICATORS FOR DISTINGUISHING ACTIVE SPEAKERS OF USERS DISPLAYED AS THREE-DIMENSIONAL REPRESENTATIONS

Simplified Explanation

The patent application describes a system that automates visual indicators to show active speakers in a communication session using 3D representations.

  • Participants in the communication session are displayed as 3D avatars within a 3D environment.
  • The user interface also includes a gallery of 2D images of other participants.
  • When a user starts speaking, the system detects their activity through audio signals from their device.
  • In response, the system automatically adds a complementary image of the user to the gallery.
  • This helps viewers navigate through complex user interface arrangements with many avatars.


Original Abstract Submitted

The disclosed techniques provide systems that automate visual indicators to show active speakers of a communication session who are displayed as 3D representations. Some participants of a communication session can be displayed in a user interface using 3D representations, e.g., avatars, that are each positioned within a 3D environment. The user interface may also include and number of renderings of 2D images of other participants displayed in a gallery, e.g., a display region that is designated for active speakers. When a user who is displayed as a 3D representation starts to speak, the system can detect the speaker's activity via a detection of an audio signal from the user's device. In response to the detection, the system can then automatically add a complementary image of the user to the gallery. The complementary image can help viewers navigate through complex user interface arrangements that display a large number of avatars.