#自然语言处理#Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
Data extraction with LLM on CPU
#自然语言处理#中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Low-bit LLM inference on CPU with lookup table
Data extraction with LLM on CPU
Data extraction with LLM on CPU
Data extraction with LLM on CPU
Easy and Fast LLM finetuning on GPU or CPU.
Production ready LLM model compression/quantization toolkit with accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.
Universal and Transferable Attacks on Aligned Language Models
the LLM vulnerability scanner
#大语言模型#LLM Finetuning with peft
#大语言模型#[ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
LLM as a Chatbot Service
#大语言模型#Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such ...
#大语言模型#LlamaIndex is a data framework for your LLM applications
Numbers every LLM developer should know
📖A curated list of Awesome LLM/VLM Inference Papers with codes, such as FlashAttention, PagedAttention, Parallelism, etc. 🎉🎉
Simple UI for LLM Model Finetuning
#计算机科学#Build LLM-powered applications in Ruby