Skip to content

chadHGY/awesome-deep-model-compression

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 

Repository files navigation

Awesome Deep Model Compression

Awesome hackmd-github-sync-badge

A useful list of Deep Model Compression related research papers, articles, tutorials, libraries, tools and more.
Currently the Repos are additional given tags either [Pytorch/TF]. To quickly find hands-on Repos in your commonly used framework, please Ctrl+F to get start 😃

Contents


Papers

General

Architecture

Quantization

Binarization

Pruning

  • The Lottery Ticket Hypothesis | ICLR, 2019, Google | Paper | Code

Distillation

Low Rank Approximation


Articles

Blogs


Tools

Libraries

  • torch.nn.utils.prune [Pytorch]
    Pytorch official supported sparsify neural networks and custom pruning technique.
  • Neural Network Intelligence[Pytorch/TF] Star on GitHub
    There are some popular model compression algorithms built-in in NNI. Users could further use NNI’s auto tuning power to find the best compressed model, which is detailed in Auto Model Compression.
  • Condensa [Pytorch] Star on GitHub
    A Programming System for Neural Network Compression. | paper
  • IntelLabs distiller [Pytorch] Star on GitHub
    Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. | Documentation
  • Torch-Pruning[Pytorch] Star on GitHub
    A pytorch toolkit for structured neural network pruning and layer dependency.
  • CompressAI [Pytorch] Star on GitHub
    A PyTorch library and evaluation platform for end-to-end compression research.
  • Model Compression[Pytorch] Star on GitHub
    A onestop pytorch model compression repo. | Reposhub
  • TensorFlow Model Optimization Toolkit [TF] Star on GitHub
    Accompanied blog post, TensorFlow Model Optimization Toolkit — Pruning API
  • XNNPACK Star on GitHub
    XNNPACK is a highly optimized library of floating-point neural network inference operators for ARM, WebAssembly, and x86 (SSE2 level) platforms. It's a based on QNNPACK library. However, unlike QNNPACK, XNNPACK focuses entirely on floating-point operators.

Cross Platform

Hard-ware Integration

  • Pytorch Glow Star on GitHub
    Glow is a machine learning compiler and execution engine for hardware accelerators. It is designed to be used as a backend for high-level machine learning frameworks. The compiler is designed to allow state of the art compiler optimizations and code generation of neural network graphs.

  • CoreML (Apple) [Pytorch,TF,Keras,SKLearn ...etc] Star on GitHub
    Core ML provides a unified representation for all models. Your app uses Core ML APIs and user data to make predictions, and to train or fine-tune models, all on the user’s device.
    Introduction