#计算机科学# RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
#计算机科学# Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
This repository contains demos I made with the Transformers library by HuggingFace.
翻译 - 这个存储库包含我用 HuggingFace 的 Transformers 库制作的演示。
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
翻译 - 使用Mesh-tensorflow库,可并行执行模型GPT2和类似GPT3的模型,并能够扩展到完整的GPT3尺寸(甚至可能更多!)。
#自然语言处理# Chinese version of GPT2 training code, using BERT tokenizer.
翻译 - 使用BERT标记程序的中文版GPT2培训代码。
#自然语言处理# Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
#大语言模型# Prompt Engineering | Prompt Versioning | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engineering, LLMs and other latest research
#自然语言处理# An unnecessarily tiny implementation of GPT-2 in NumPy.
#自然语言处理# GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
#大语言模型# Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJ...
#计算机科学# Large-scale pretraining for dialogue
翻译 - 对话的大规模预培训
#计算机科学# Guide to using pre-trained large language models of source code
#自然语言处理# GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
This Discord chatbot is incredibly versatile. Powered incredibly fast Groq API
TextBox 2.0 is a text generation library with pre-trained language models
翻译 - TextBox是用于构建文本生成系统的开源库。
#大语言模型# This is a repository that aims to provide updates on the status of jailbreaking the OpenAI GPT language model.
Code and data for crosstalk text generation tasks, exploring whether large models and pre-trained language models can understand humor.