Skip to content
#

quantization-aware-training

Here are 54 public repositories matching this topic...

Our work implements novel L2-Norm gradient (L2Grad) and variance of the weight distrbution (VarianceNorm) regularizers for quantization-aware training such that the distribution of weights are more compatible with post-training quantization especially for low bit-widths. We provide a theoretical basis that directly relates L2-Grad with post quan…

  • Updated May 15, 2021
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the quantization-aware-training topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the quantization-aware-training topic, visit your repo's landing page and select "manage topics."

Learn more