Intel corporation (20240184572). INSTRUCTIONS AND LOGIC TO PERFORM FLOATING POINT AND INTEGER OPERATIONS FOR MACHINE LEARNING simplified abstract

From WikiPatents
Jump to navigation Jump to search

INSTRUCTIONS AND LOGIC TO PERFORM FLOATING POINT AND INTEGER OPERATIONS FOR MACHINE LEARNING

Organization Name

intel corporation

Inventor(s)

Himanshu Kaul of Portland OR (US)

Mark A. Anders of Hillsboro OR (US)

Sanu K. Mathew of Hillsboro OR (US)

Anbang Yao of Beijing (CN)

Joydeep Ray of Folsom CA (US)

Ping T. Tang of Edison NJ (US)

Michael S. Strickland of Sunnyvale CA (US)

Xiaoming Chen of Shanghai (CN)

Tatiana Shpeisman of Menlo Park CA (US)

Abhishek R. Appu of El Dorado Hills CA (US)

Altug Koker of El Dorado Hills CA (US)

Kamal Sinha of Rancho Cordova CA (US)

Balaji Vembu of Folsom CA (US)

Nicolas C. Galoppo Von Borries of Portland OR (US)

Eriko Nurvitadhi of Hillsboro OR (US)

Rajkishore Barik of Santa Clara CA (US)

Tsung-Han Lin of Campbell CA (US)

Vasanth Ranganathan of El Dorado Hills CA (US)

Sanjeev Jahagirdar of Folsom CA (US)

INSTRUCTIONS AND LOGIC TO PERFORM FLOATING POINT AND INTEGER OPERATIONS FOR MACHINE LEARNING - A simplified explanation of the abstract

This abstract first appeared for US patent application 20240184572 titled 'INSTRUCTIONS AND LOGIC TO PERFORM FLOATING POINT AND INTEGER OPERATIONS FOR MACHINE LEARNING

Simplified Explanation

The abstract describes a graphics processing unit that accelerates machine-learning operations by performing a two-dimensional matrix multiply and accumulate operation.

  • Graphics processing unit with a single instruction, multiple thread (SIMT) architecture
  • Includes a multiprocessor that executes at least one single instruction
  • Contains a first compute unit within the multiprocessor to perform the matrix operation
  • The operation involves computing an intermediate product of 16-bit operands and a 32-bit sum based on the product

Potential Applications

This technology can be applied in various fields such as artificial intelligence, image and video processing, scientific computing, and data analytics.

Problems Solved

This technology solves the problem of accelerating machine-learning operations and improving the efficiency of two-dimensional matrix operations.

Benefits

The benefits of this technology include faster processing speeds, increased performance in machine-learning tasks, and enhanced capabilities for complex mathematical computations.

Potential Commercial Applications

Potential commercial applications of this technology include in data centers, cloud computing services, autonomous vehicles, robotics, and healthcare for medical imaging analysis.

Possible Prior Art

One possible prior art could be the use of general-purpose graphics processing units (GPGPUs) for accelerating machine-learning operations. These GPGPUs have been used in various applications requiring parallel processing capabilities.

What are the energy efficiency implications of using this technology?

Using this technology can lead to improved energy efficiency due to the accelerated processing speeds, which can reduce the overall power consumption of computing systems.

How does this technology compare to other hardware accelerators for machine learning?

This technology offers a specialized approach to accelerating machine-learning operations specifically tailored for two-dimensional matrix operations, providing a unique advantage over general-purpose hardware accelerators.


Original Abstract Submitted

one embodiment provides for a graphics processing unit to accelerate machine-learning operations, the graphics processing unit comprising a multiprocessor having a single instruction, multiple thread (simt) architecture, the multiprocessor to execute at least one single instruction; and a first compute unit included within the multiprocessor, the at least one single instruction to cause the first compute unit to perform a two-dimensional matrix multiply and accumulate operation, wherein to perform the two-dimensional matrix multiply and accumulate operation includes to compute an intermediate product of 16-bit operands and to compute a 32-bit sum based on the intermediate product.