#大语言模型#LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
#数据仓库#Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
#计算机科学#An Open-sourced Knowledgable Large Language Model Framework.
#大语言模型#Large Language Models for All, 🦙 Cult and More, Stay in touch !
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train...
Best practices & guides on how to write distributed pytorch training code
#大语言模型#Super-Efficient RLHF Training of LLMs with Parameter Reallocation
llama2 finetuning with deepspeed and lora
#大语言模型#A full pipeline to finetune ChatGLM LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the ChatGLM architecture. Basically ChatG...
#自然语言处理#Train llm (bloom, llama, baichuan2-7b, chatglm3-6b) with deepspeed pipeline mode. Faster than zero/zero++/fsdp.
llm-inference is a platform for publishing and managing llm inference, providing a wide range of out-of-the-box features for model deployment, such as UI, RESTful API, auto-scaling, computing resource...
Scripts of LLM pre-training and fine-tuning (w/wo LoRA, DeepSpeed)