Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
-
Updated
May 8, 2024 - Python
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Code for paper " AdderNet: Do We Really Need Multiplications in Deep Learning?"
Deep Face Model Compression
EfficientFormerV2 [ICCV 2023] & EfficientFormer [NeurIPs 2022]
[ICML 2024] LLMCompiler: An LLM Compiler for Parallel Function Calling
Learning Efficient Convolutional Networks through Network Slimming, In ICCV 2017.
"LightGaussian: Unbounded 3D Gaussian Compression with 15x Reduction and 200+ FPS", Zhiwen Fan, Kevin Wang, Kairun Wen, Zehao Zhu, Dejia Xu, Zhangyang Wang
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
List of papers related to neural network quantization in recent AI conferences and journals.
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
[CVPR 2021] Exploring Sparsity in Image Super-Resolution for Efficient Inference
[ICLR 2022] Code for Graph-less Neural Networks: Teaching Old MLPs New Tricks via Distillation (GLNN)
Code for paper 'Multi-Component Optimization and Efficient Deployment of Neural-Networks on Resource-Constrained IoT Hardware'
(CVPR 2021, Oral) Dynamic Slimmable Network
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
[ECCV2022] Efficient Long-Range Attention Network for Image Super-resolution
[ECCV 2022] Official implementation of the paper "DeciWatch: A Simple Baseline for 10x Efficient 2D and 3D Pose Estimation"
Explorations into some recent techniques surrounding speculative decoding
Code for WF-IoT paper 'TinyML Benchmark: Executing Fully Connected Neural Networks on Commodity Microcontrollers'
Soft Threshold Weight Reparameterization for Learnable Sparsity
Add a description, image, and links to the efficient-inference topic page so that developers can more easily learn about it.
To associate your repository with the efficient-inference topic, visit your repo's landing page and select "manage topics."