Low-Rank Compression of Neural Networks

Number of patents in Portfolio can not be more than 2000

United States of America

APP PUB NO 20250021826A1
SERIAL NO

18629162

Stats

ATTORNEY / AGENT: (SPONSORED)

Importance

Loading Importance Indicators... loading....

Abstract

See full text

In one embodiment, a method includes accessing at least a portion of a training dataset for a trained neural network that includes multiple layers, where each layer includes a number of parameters, and where the training dataset includes multiple training samples that each include an input and a ground-truth output used to train the trained neural network. The method further includes training a hypernetwork to generate a layer-specific compression mask for each of one or more of the multiple layers of the trained neural network. The method further includes generating, by the trained hypernetwork, a final layer-specific compression mask for the trained neural network and compressing the trained neural network by reducing, for each of the one or more layers of the neural network, the number of parameters of that layer according to the final layer-specific compression mask.

Loading the Abstract Image... loading....

First Claim

See full text

Family

Loading Family data... loading....

Patent Owner(s)

Patent OwnerAddress
SAMSUNG ELECTRONICS CO LTDSUWON-SI

International Classification(s)

  • [Classification Symbol]
  • [Patents Count]

Inventor(s)

Inventor Name Address # of filed Patents Total Citations
Gao, Shangqian Mountain View, US 14 32
Hsu, Yen-Chang Fremont, US 11 1
Hua, Ting Cupertino, US 18 7
Jin, Hongxia San Jose, US 149 1846
Shen, Yilin Mountain View, US 62 197

Cited Art Landscape

Load Citation

Patent Citation Ranking

Forward Cite Landscape

Load Citation