Qualcomm incorporated (20240095542). SPLIT NEURAL NETWORK ACCELERATION ARCHITECTURE SCHEDULING AND DYNAMIC INFERENCE ROUTING simplified abstract

From WikiPatents
Jump to navigation Jump to search

SPLIT NEURAL NETWORK ACCELERATION ARCHITECTURE SCHEDULING AND DYNAMIC INFERENCE ROUTING

Organization Name

qualcomm incorporated

Inventor(s)

Vijaya Kumar Kilari of Bangalore (IN)

Fan Wu of Broomfield CO (US)

Geoffrey Carlton Berry of Durham NC (US)

Hemanth Puranik of Bangalore (IN)

SPLIT NEURAL NETWORK ACCELERATION ARCHITECTURE SCHEDULING AND DYNAMIC INFERENCE ROUTING - A simplified explanation of the abstract

This abstract first appeared for US patent application 20240095542 titled 'SPLIT NEURAL NETWORK ACCELERATION ARCHITECTURE SCHEDULING AND DYNAMIC INFERENCE ROUTING

Simplified Explanation

The abstract describes a method for accelerating machine learning on a computing device by splitting a neural network into sub-neural networks and hosting them in inference accelerators.

  • Accessing a neural network
  • Splitting the neural network into n sub-neural networks
  • Hosting the n sub-neural networks in m inference accelerators
  • Scheduling the n sub-neural networks in the m inference accelerators
  • Executing the n sub-neural networks in the m inference accelerators

Potential Applications

This technology could be applied in various fields such as autonomous vehicles, healthcare diagnostics, natural language processing, and image recognition.

Problems Solved

This technology addresses the challenge of optimizing machine learning processes on computing devices by efficiently distributing the workload among multiple inference accelerators.

Benefits

The method allows for faster and more efficient execution of machine learning tasks, leading to improved performance and reduced processing time.

Potential Commercial Applications

  • "Accelerating Machine Learning on Computing Devices: Enhancing Performance and Efficiency"

Possible Prior Art

One possible prior art could be the use of parallel processing techniques in machine learning to improve performance and speed of computations.

Unanswered Questions

How does this method compare to other techniques for accelerating machine learning tasks?

The article does not provide a comparison with other methods or technologies in the field.

What are the limitations or constraints of implementing this method in real-world applications?

The article does not address any potential challenges or limitations that may arise when implementing this method in practical scenarios.


Original Abstract Submitted

a method for accelerating machine learning on a computing device is described. the method includes accessing a neural network. the method also includes splitting the neural network into n sub-neural networks. the method further includes hosting the n sub-neural networks in m inference accelerators. the method also includes scheduling the n sub-neural networks in the m inference accelerators. the method further includes executing the n sub-neural networks in the m inference accelerators.