SONY INTERACTIVE ENTERTAINMENT INC. (20240284011). APPARATUS AND METHODS FOR CONTENT DESCRIPTION simplified abstract

From WikiPatents
Jump to navigation Jump to search

APPARATUS AND METHODS FOR CONTENT DESCRIPTION

Organization Name

SONY INTERACTIVE ENTERTAINMENT INC.

Inventor(s)

Ryan Spick of London (GB)

Timothy Edward Bradley of London (GB)

Guy David Moss of London (GB)

Ayush Raina of London (GB)

Pierluigi Amadori of London (GB)

APPARATUS AND METHODS FOR CONTENT DESCRIPTION - A simplified explanation of the abstract

This abstract first appeared for US patent application 20240284011 titled 'APPARATUS AND METHODS FOR CONTENT DESCRIPTION

The abstract describes a data processing apparatus that determines description data for content using a video captioning model trained to detect motions of animated objects in video images and generate captions based on these motions.

  • Video captioning model trained to detect predetermined motions of animated objects in video images
  • Captions generated based on detected motions, comprising caption data with words describing the motions
  • Caption data includes audio, text, and image data
  • Output circuitry provides description data based on the generated captions

Potential Applications: - Automated video content description for accessibility purposes - Enhancing searchability of video content through detailed descriptions - Improving user experience by providing informative captions for video content

Problems Solved: - Lack of detailed descriptions for video content - Accessibility issues for individuals with hearing impairments - Inefficient manual generation of captions for videos

Benefits: - Enhanced accessibility for a wider audience - Improved search engine optimization for video content - Streamlined process for generating captions and descriptions

Commercial Applications: Title: Automated Video Captioning and Description Generation Technology This technology can be utilized in: - Video streaming platforms - E-learning platforms - Digital marketing agencies

Questions about the Technology: 1. How does the video captioning model determine the motions of animated objects in video images? 2. What are the different types of data included in the caption data generated by the model?

Frequently Updated Research: Stay updated on advancements in video captioning technology and accessibility tools to enhance the capabilities of the data processing apparatus.


Original Abstract Submitted

a data processing apparatus for determining description data for describing content includes: a video captioning model to receive an input comprising at least video images associated with the content, wherein the video captioning model is trained to detect one or more predetermined motions of one or more animated objects in the video images and determine one or more captions in dependence on one or more of the predetermined motions, one or more of the captions comprising respective caption data comprising one or more words for describing one or more of the predetermined motions, the respective caption data comprising one or more of audio data, text data and image data; and output circuitry to output description data in dependence on one or more of the captions.