GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

quantization

Website
Wikipedia
hiyouga/LLaMA-Factory
https://static.github-zh.com/github_avatars/hiyouga?size=40
hiyouga / LLaMA-Factory

#自然语言处理#Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

fine-tuningllama大语言模型pefttransformersrlhfqloraquantizationqweninstruction-tuninggptloralarge-language-modelsagent人工智能moellama3deepseekgemma自然语言处理
Python 54.23 k
7 小时前
ymcui/Chinese-LLaMA-Alpaca
https://static.github-zh.com/github_avatars/ymcui?size=40
ymcui / Chinese-LLaMA-Alpaca

#自然语言处理#中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

大语言模型plmpre-trained-language-modelsalpacallama自然语言处理quantizationlarge-language-modelsloraalpaca-2llama-2
Python 18.87 k
1 年前
https://static.github-zh.com/github_avatars/SYSTRAN?size=40
SYSTRAN / faster-whisper

#计算机科学#Faster Whisper transcription with CTranslate2

深度学习inferencequantizationspeech-recognitionspeech-to-texttransformerWhisperopenai
Python 17.04 k
1 个月前
UFund-Me/Qbot
https://static.github-zh.com/github_avatars/UFund-Me?size=40
UFund-Me / Qbot

#区块链#[🔥updating ...] AI 自动量化交易机器人(完全本地部署) AI-powered Quantitative Investment Research Platform. 📃 online docs: https://ufund-me.github.io/Qbot ✨ :news: qbot-mini: https://github.com/Charmve/iQuant

funds机器学习pytradequantitative-financequantitative-tradingquantizationstrategiestrademarksquant-trader比特币区块链深度学习fintechbacktest
Jupyter Notebook 12.2 k
8 天前
bitsandbytes-foundation/bitsandbytes
https://static.github-zh.com/github_avatars/bitsandbytes-foundation?size=40
bitsandbytes-foundation / bitsandbytes

#大语言模型#Accessible large language models via k-bit quantization for PyTorch.

大语言模型机器学习PyTorchqloraquantization
Python 7.22 k
6 天前
https://static.github-zh.com/github_avatars/kornelski?size=40
kornelski / pngquant

Lossy PNG compressor — pngquant command based on libimagequant library

pngquantCode qualitypngpng-compressionquantizationstdinpaletteconversionimage-optimizationC
C 5.41 k
7 天前
https://static.github-zh.com/github_avatars/AutoGPTQ?size=40
AutoGPTQ / AutoGPTQ

#自然语言处理#An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.

transformers深度学习inferencelarge-language-models大语言模型自然语言处理PyTorchquantizationtransformer
Python 4.89 k
3 个月前
https://static.github-zh.com/github_avatars/IntelLabs?size=40
IntelLabs / distiller

Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller

PyTorchpruningquantizationJupyter Notebook深度神经网络regularizationdistillationonnx
Jupyter Notebook 4.4 k
2 年前
https://static.github-zh.com/github_avatars/OpenNMT?size=40
OpenNMT / CTranslate2

#计算机科学#Fast inference engine for Transformer models

neural-machine-translationC++mklquantizationCUDAthrustopennmt深度神经网络openmponednnintrinsicsavx2avxparallel-computinggemmneontransformer-modelsmachine-translation深度学习inference
C++ 3.91 k
3 个月前
neuralmagic/deepsparse
https://static.github-zh.com/github_avatars/neuralmagic?size=40
neuralmagic / deepsparse

#自然语言处理#Sparsity-aware deep learning inference runtime for CPUs

机器学习onnxinference机器视觉object-detectionpruningquantizationpretrained-models自然语言处理cpussparsificationllm-inferenceperformance
Python 3.16 k
1 个月前
https://static.github-zh.com/github_avatars/huawei-noah?size=40
huawei-noah / Pretrained-Language-Model

Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

knowledge-distillationmodel-compressionquantizationpretrained-models
Python 3.11 k
1 年前
https://static.github-zh.com/github_avatars/huggingface?size=40
huggingface / optimum

🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools

onnxPyTorchinferencetrainingintelgraphcoreonnxruntimetransformersquantizationhabanaoptimizationtflite
Python 2.98 k
3 天前
IntelLabs/nlp-architect
https://static.github-zh.com/github_avatars/IntelLabs?size=40
IntelLabs / nlp-architect

#自然语言处理#A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks

深度学习自然语言处理nluTensorflowdynetPyTorchberttransformersquantization
Python 2.94 k
3 年前
https://static.github-zh.com/github_avatars/aaron-xichen?size=40
aaron-xichen / pytorch-playground

Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)

pytorch-tutorialpytorch-tutorialsPyTorchquantization
Python 2.68 k
3 年前
https://static.github-zh.com/github_avatars/stochasticai?size=40
stochasticai / xTuring

#大语言模型#Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJ...

深度学习fine-tuninggpt-2gpt-jllama大语言模型loralanguage-modelalpacafinetuningadaptergen-aigenerative-aimistralpeftquantization
Python 2.66 k
10 个月前
https://static.github-zh.com/github_avatars/intel?size=40
intel / neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

low-precisionpruningsparsityauto-tuningknowledge-distillationquantizationquantization-aware-trainingpost-training-quantizationsmoothquantlarge-language-modelsgptqint8
Python 2.45 k
12 小时前
https://static.github-zh.com/github_avatars/mit-han-lab?size=40
mit-han-lab / nunchaku

[ICLR2025 Spotlight] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models

diffusion-modelsfluxgenailoramlsysquantizationiclriclr2025comfyui
Python 2.37 k
17 小时前
https://static.github-zh.com/github_avatars/quic?size=40
quic / aimet

#计算机科学#AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.

quantization深度学习compressionOpen Source机器学习pruningauto-ml深度神经网络
Python 2.37 k
2 天前
https://static.github-zh.com/github_avatars/dvmazur?size=40
dvmazur / mixtral-offloading

#大语言模型#Run Mixtral-8x7B models in Colab or consumer desktops

colab-notebook深度学习google-colablanguage-model大语言模型mixture-of-expertsoffloadingPyTorchquantization
Python 2.31 k
1 年前
https://static.github-zh.com/github_avatars/666DZY666?size=40
666DZY666 / micronet

micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Ari...

quantizationpruningdorefatwnbnnxnor-netPyTorchmodel-compressiongroup-convolutionconvolutional-networksquantization-aware-trainingpost-training-quantizationtensorrtonnx
Python 2.25 k
2 个月前
loading...