Simpler Distil-Whisper
-
Updated
May 30, 2024 - Python
Simpler Distil-Whisper
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
On-device LLM Inference Powered by X-Bit Quantization
Characterization study repository for model compression method: pruning
List of papers related to neural network quantization in recent AI conferences and journals.
Gather research papers, corresponding codes (if having), reading notes and any other related materials about Hot🔥🔥🔥 fields in Computer Vision based on Deep Learning.
This repo contains model compression(using TensorRT) and documentation of running various deep learning models on NVIDIA Jetson Orin, Nano (aarch64 architectures)
记录有意思的AI相关项目
ptdeco is a library for model optimization by decomposition built on top of PyTorch
Awesome Knowledge Distillation
A curated list of awesome NLP, Computer Vision, Model Compression, XAI, Reinforcement Learning, Security etc Paper
Dive into advanced quantization techniques. Learn to implement and customize linear quantization functions, measure quantization error, and compress model weights using PyTorch for efficient and accessible AI models.
Hyperparameter Tuning with Microsoft NNI to automated machine learning (AutoML) experiments. The tool dispatches and runs trial jobs generated by tuning algorithms to search the best neural architecture and/or hyper-parameters in different environments like local machine, remote servers and cloud.
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
Resources of our survey paper "Enabling AI on Edges: Techniques, Applications and Challenges"
模型压缩的小白入门教程
a collection of computer vision projects&tools. 计算机视觉方向项目和工具集合。
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
OpenBA-V2: 3B LLM (Large Language Model) with T5 architecture, utilizing model pruning technique and continuing pretraining from OpenBA-15B.
Add a description, image, and links to the model-compression topic page so that developers can more easily learn about it.
To associate your repository with the model-compression topic, visit your repo's landing page and select "manage topics."