Energy Efficient Computations of Attention-based Inferences

Number of patents in Portfolio can not be more than 2000

United States of America Patent

APP PUB NO 20240281428A1
SERIAL NO

18414782

Stats

ATTORNEY / AGENT: (SPONSORED)

Importance

Loading Importance Indicators... loading....

Abstract

See full text

An apparatus to compute an attention matrix implementing an attention mechanism in artificial neural networks, having: memory to store key value pairs; a reorder buffer to provide a reordered list of keys from the key value pairs; an analog dot product accelerator configured to compute dot products of key elements of keys from the reordered list of keys with respective query elements of a query row of a query matrix; a processing device configured to generate, based on results of the dot products, a row of attention scores corresponding to the query row of the query matrix for the reordered list of keys; and a further accelerator configured to compute dot products of segments of the attention scores with value elements of respective segments of values from a list of values from the key value pairs to generate an attention matrix.

Loading the Abstract Image... loading....

First Claim

See full text

Family

Loading Family data... loading....

Patent Owner(s)

Patent OwnerAddress
MICRON TECHNOLOGY INC8000 SOUTH FEDERAL WAY MAILSTOP 1-507 BOISE ID 83707-0006

International Classification(s)

  • [Classification Symbol]
  • [Patents Count]

Inventor(s)

Inventor Name Address # of filed Patents Total Citations
Kale, Poorna Folsom, US 246 1287
Lakshman, Shashank Bangalore Folsom, US 7 0
Sunny, Febin Folsom, US 20 0
Tiku, Saideep Folsom, US 52 12

Cited Art Landscape

Load Citation

Patent Citation Ranking

Forward Cite Landscape

Load Citation