#自然语言处理# MOSS 是复旦大学开源的一个支持中英双语和多种插件的对话语言模型
#自然语言处理# Chinese version of GPT2 training code, using BERT tokenizer.
翻译 - 使用BERT标记程序的中文版GPT2培训代码。
Python package to easily retrain OpenAI's GPT-2 text-generating model on new texts
翻译 - Python软件包可轻松地在新文本上重新训练OpenAI的GPT-2文本生成模型
#自然语言处理# GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
#计算机科学# Large-scale pretraining for dialogue
翻译 - 对话的大规模预培训
#自然语言处理# A modular RL library to fine-tune language models to human preferences
#自然语言处理# ⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
#自然语言处理# GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
TextBox 2.0 is a text generation library with pre-trained language models
翻译 - TextBox是用于构建文本生成系统的开源库。
#自然语言处理# Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Transformers-based text generation models.
#计算机科学# Large-scale pretrained models for goal-directed dialog
#自然语言处理# We introduced a new model designed for the Code generation task. Its test accuracy on the HumanEval base dataset surpasses that of GPT-4 Turbo (April 2024) and GPT-4o.
#大语言模型# The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
#自然语言处理# [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
Code and data for crosstalk text generation tasks, exploring whether large models and pre-trained language models can understand humor.