17525999. PARAMETER REDUNDANCY REDUCTION METHOD simplified abstract (International Business Machines Corporation)

From WikiPatents
Jump to navigation Jump to search

PARAMETER REDUNDANCY REDUCTION METHOD

Organization Name

International Business Machines Corporation

Inventor(s)

Zhong Fang Yuan of Xi'an (CN)

Tong Liu of Xi'an (CN)

Li Juan Gao of Xi'an (CN)

Na Liu of Xi'an (CN)

Xiang Yu Yang of Xi'an (CN)

PARAMETER REDUNDANCY REDUCTION METHOD - A simplified explanation of the abstract

This abstract first appeared for US patent application 17525999 titled 'PARAMETER REDUNDANCY REDUCTION METHOD

Simplified Explanation

The patent application describes a method, computer program product, and computer system for generating and using a basic state layer. This involves training multiple task models on a pre-trained backbone model, where each task model has different parameter matrices in its feature layers. An encoder-decoder model is then trained, which compresses the parameter matrices into hidden layers to generate a basic state model. The decoder part of the model generates an output layer identical to the input layer.

  • The patent application describes a method for generating and using a basic state layer.
  • Multiple task models are trained on a pre-trained backbone model.
  • Each task model has different parameter matrices in its feature layers.
  • An encoder-decoder model is trained, which compresses the parameter matrices into hidden layers.
  • The encoder part of the model maps and compresses the parameter matrices into hidden layers to generate a basic state model.
  • The decoder part of the model generates an output layer identical to the input layer.

Potential Applications

  • This technology can be applied in various fields where multiple task models need to be trained and used.
  • It can be used in machine learning applications for tasks such as image recognition, natural language processing, and speech recognition.
  • The basic state layer generated by this method can be used as a representation of the input data for further processing or analysis.

Problems Solved

  • This technology solves the problem of training multiple task models with different parameter matrices.
  • It provides a method to compress and represent the parameter matrices into a basic state model.
  • The use of an encoder-decoder model allows for efficient mapping and generation of the output layer.

Benefits

  • The method allows for efficient training and utilization of multiple task models.
  • It provides a compressed representation of the input data through the basic state model.
  • The use of an encoder-decoder model enables the generation of an output layer identical to the input layer.


Original Abstract Submitted

A method, computer program product, and computer system for generating and using a basic state layer. N task models are provided (N ≥ 2). Each task model was trained on a same pre-trained backbone model. Each task model includes M feature layers and a task layer (M ≥ 1). Each feature layer of each task model includes a parameter matrix that is different for the different models. An encoder-decoder model is trained. The encoder-decoder model includes sequentially: an input layer, an encoder, M hidden layers, a decoder, and an output layer. The encoder is a neural network that maps and compresses the parameter matrices in the input layer into the M hidden layers, which generates a basic state model. The decoder is a neural network that receives the basic state model as input and generates the output layer to be identical to the input layer.