#计算机科学#RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
#计算机科学#Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
This repository contains demos I made with the Transformers library by HuggingFace.
翻译 - 这个存储库包含我用 HuggingFace 的 Transformers 库制作的演示。
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
翻译 - 使用Mesh-tensorflow库,可并行执行模型GPT2和类似GPT3的模型,并能够扩展到完整的GPT3尺寸(甚至可能更多!)。
#自然语言处理#Chinese version of GPT2 training code, using BERT tokenizer.
翻译 - 使用BERT标记程序的中文版GPT2培训代码。
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
#自然语言处理#Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
#大语言模型#Prompt Engineering | Prompt Versioning | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engineering, LLMs and other latest research
#自然语言处理#GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
#大语言模型#Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJ...
#计算机科学#Large-scale pretraining for dialogue
翻译 - 对话的大规模预培训
#计算机科学#Guide to using pre-trained large language models of source code
#自然语言处理#GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
This Discord chatbot is incredibly versatile. Powered incredibly fast Groq API
TextBox 2.0 is a text generation library with pre-trained language models
翻译 - TextBox是用于构建文本生成系统的开源库。
#大语言模型#This is a repository that aims to provide updates on the status of jailbreaking the OpenAI GPT language model.
Code and data for crosstalk text generation tasks, exploring whether large models and pre-trained language models can understand humor.