20240056761. THREE-DIMENSIONAL (3D) SOUND RENDERING WITH MULTI-CHANNEL AUDIO BASED ON MONO AUDIO INPUT simplified abstract (SAMSUNG ELECTRONICS CO., LTD.)

From WikiPatents
Jump to navigation Jump to search

THREE-DIMENSIONAL (3D) SOUND RENDERING WITH MULTI-CHANNEL AUDIO BASED ON MONO AUDIO INPUT

Organization Name

SAMSUNG ELECTRONICS CO., LTD.

Inventor(s)

Vijendra Raj Apsingekar of San Jose CA (US)

Akash Sahoo of San Jose CA (US)

Anil S. Yadav of San Jose CA (US)

Sivakumar Balasubramanian of Sunnyvale CA (US)

THREE-DIMENSIONAL (3D) SOUND RENDERING WITH MULTI-CHANNEL AUDIO BASED ON MONO AUDIO INPUT - A simplified explanation of the abstract

This abstract first appeared for US patent application 20240056761 titled 'THREE-DIMENSIONAL (3D) SOUND RENDERING WITH MULTI-CHANNEL AUDIO BASED ON MONO AUDIO INPUT

Simplified Explanation

The patent application describes a method for analyzing video content and associated audio content. The method involves detecting objects in the video, determining their position or motion trajectory, and classifying them into different object classes. The audio content is separated into multiple audio streams based on the video content, and each audio stream is associated with a specific audio source. The audio sources are also classified into the same object classes as the detected objects. Finally, the audio streams associated with the same object class are distributed into multiple audio channels based on the position or motion trajectory of the corresponding object.

  • The method obtains video content and substantially mono audio content.
  • It determines the position or motion trajectory of detected objects in the video.
  • It classifies the objects into multiple object classes.
  • The audio content is separated into audio streams based on the video content.
  • Each audio stream is associated with a specific audio source.
  • The audio sources are classified into the same object classes as the detected objects.
  • The audio streams associated with the same object class are distributed into multiple audio channels based on the position or motion trajectory of the corresponding object.

Potential Applications:

  • Enhancing video content by separating and distributing audio streams based on the detected objects.
  • Improving audio quality and spatial representation in video content.
  • Enabling immersive audio experiences in virtual reality or augmented reality applications.

Problems Solved:

  • Efficiently analyzing and processing video and audio content together.
  • Enhancing the synchronization between audio and video elements in multimedia applications.
  • Providing a more immersive and realistic audiovisual experience.

Benefits:

  • Improved audio quality and spatial representation in video content.
  • Enhanced user experience in multimedia applications.
  • Increased realism and immersion in virtual reality or augmented reality environments.


Original Abstract Submitted

a method includes obtaining video content and associated substantially mono audio content. the method also includes determining at least one of a position or a motion trajectory of each of one or more objects detected in the video content and classifying each of the one or more objects into one of multiple object classes. the method further includes separating audio streams within the audio content based on the video content. each of the audio streams is associated with one of multiple audio sources. the method also includes classifying each of the audio sources into one of the object classes. in addition, the method includes, for each audio source classified into the same object class as one of the one or more objects, distributing the audio stream associated with that audio source into multiple audio channels based on at least one of the position or the motion trajectory of that object.