Compression of Machine-Learned Models via Entropy Penalized Weight Reparameterization

Number of patents in Portfolio can not be more than 2000

United States of America Patent

APP PUB NO 20240220863A1
SERIAL NO

18409520

Stats

ATTORNEY / AGENT: (SPONSORED)

Importance

Loading Importance Indicators... loading....

Abstract

See full text

Example aspects of the present disclosure are directed to systems and methods that learn a compressed representation of a machine-learned model (e.g., neural network) via representation of the model parameters within a reparameterization space during training of the model. In particular, the present disclosure describes an end-to-end model weight compression approach that employs a latent-variable data compression method. The model parameters (e.g., weights and biases) are represented in a “latent” or “reparameterization” space, amounting to a reparameterization. In some implementations, this space can be equipped with a learned probability model, which is used first to impose an entropy penalty on the parameter representation during training, and second to compress the representation using arithmetic coding after training. The proposed approach can thus maximize accuracy and model compressibility jointly, in an end-to-end fashion, with the rate-error trade-off specified by a hyperparameter.

Loading the Abstract Image... loading....

First Claim

See full text

Family

Loading Family data... loading....

Patent Owner(s)

Patent OwnerAddress
GOOGLE LLC1600 AMPHITHEATRE PARKWAY MOUNTAIN VIEW CA 94043

International Classification(s)

  • [Classification Symbol]
  • [Patents Count]

Inventor(s)

Inventor Name Address # of filed Patents Total Citations
Balle, Johannes San Francisco, US 11 53
Oktay, Deniz Mountain View, US 3 8
Shrivastava, Abhinav Silver Springs, US 25 2667
Singh, Saurabh Mountain View, US 101 583

Cited Art Landscape

Load Citation

Patent Citation Ranking

Forward Cite Landscape

Load Citation