Leveraging Redundancy in Attention with Reuse Transformers

Number of patents in Portfolio can not be more than 2000

United States of America Patent

APP PUB NO 20230112862A1
SERIAL NO

17960380

Stats

ATTORNEY / AGENT: (SPONSORED)

Importance

Loading Importance Indicators... loading....

Abstract

See full text

Provided are systems and methods that improve the computational efficiency of Transformers or other attention-based neural networks or machine learning models by re-using a number of attention scores between layers and/or heads of the model. To reduce the computational cost of self-attention-based models while achieving comparable or even superior results, example aspects of the present disclosure propose a novel architecture that reuses attention scores computed in one layer in one or multiple subsequent layers.

Loading the Abstract Image... loading....

First Claim

See full text

Family

Loading Family data... loading....

Patent Owner(s)

Patent OwnerAddress
GOOGLE LLC1600 AMPHITHEATRE PARKWAY MOUNTAIN VIEW CA 94043

International Classification(s)

  • [Classification Symbol]
  • [Patents Count]

Inventor(s)

Inventor Name Address # of filed Patents Total Citations
Bhojanapalli, Venkata S New York, US 1 0
Chakrabarti, Ayan New York, US 2 35
Chang, Yin-Wen New York, US 2 15
Jain, Himanshu New York, US 59 230
Kumar, Sanjiv Jericho, US 137 1460
Liu, Frederick Bellevue, US 17 1064
Lukasik, Michal New York, US 1 0
Veit, Andreas New York, US 27 127

Cited Art Landscape

Load Citation

Patent Citation Ranking

Forward Cite Landscape

Load Citation