#大语言模型#Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
#自然语言处理#中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
#计算机科学#Faster Whisper transcription with CTranslate2
#区块链#[🔥updating ...] AI 自动量化交易机器人(完全本地部署) AI-powered Quantitative Investment Research Platform. 📃 online docs: https://ufund-me.github.io/Qbot ✨ :news: qbot-mini: https://github.com/Charmve/iQuant
#大语言模型#Accessible large language models via k-bit quantization for PyTorch.
Lossy PNG compressor — pngquant command based on libimagequant library
翻译 - 有损PNG压缩器—基于libimagequant库的pngquant命令
#自然语言处理#An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
#计算机科学#Fast inference engine for Transformer models
#自然语言处理#Sparsity-aware deep learning inference runtime for CPUs
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
翻译 - 华为诺亚方舟实验室开发的预训练语言模型及其相关优化技术。
#自然语言处理#A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
翻译 - 一个模型库,用于探索最新的深度学习拓扑和优化自然语言处理神经网络的技术
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
#大语言模型#Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJ...
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
#大语言模型#Run Mixtral-8x7B models in Colab or consumer desktops
#计算机科学#AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
翻译 - AIMET是一个为经过训练的神经网络模型提供高级量化和压缩技术的库。
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Ari...
翻译 - 基于pytorch的模型压缩(1,量化:8/4 / 2bits(dorefa),三进制/二进制值(twn / bnn / xnornet); 2,修剪:常规,常规和组卷积通道修剪; 3,组卷积结构; 4,特征(A)的二进制值的分批归一化折叠)
#计算机科学#A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...