Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
-
Updated
May 9, 2024 - Python
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
Tagsy, your friendly Discord bot, designed to enhance server interaction with its intuitive tagging system
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models".
A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.
[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment
[CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
JavaScript Calculator
Explore image transformations with DeepDream Algorithm and Neural Style Transfer in creative image processing.
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
The semantic segmentation of remote sensing images
[ICLR 2019] ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware
[KDD'22] Learned Token Pruning for Transformers
[ICML'21 Oral] I-BERT: Integer-only BERT Quantization
Melanoma Classification using Semi-supervised learning
Implementation of efficient backbones for computer vision task.
[MICCAI 2021] BiX-NAS: Searching Efficient Bi-directional Architecture for Medical Image Segmentation
[ICASSP'22] Integer-only Zero-shot Quantization for Efficient Speech Recognition
S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural Networks via Guided Distribution Calibration (CVPR 2021)
Add a description, image, and links to the efficient-model topic page so that developers can more easily learn about it.
To associate your repository with the efficient-model topic, visit your repo's landing page and select "manage topics."