#大语言模型#Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
#大语言模型#LLaVA是一个具有 GPT-4V 级别功能的大语言和视觉模型助手
✨✨Latest Advances on Multimodal Large Language Models
Aligning pretrained language models with instruction data generated by themselves.
#大语言模型#Instruction Tuning with GPT-4
#自然语言处理#Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
#大语言模型#Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model
#大语言模型#🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
【EMNLP 2024🔥】Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
#大语言模型#InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions
#大语言模型#We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to...
#大语言模型#mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
[ECCV2024] Video Foundation Models & Data for Multimodal Understanding
#计算机科学#An Open-sourced Knowledgable Large Language Model Framework.
#数据仓库#A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
#自然语言处理#DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
#自然语言处理#Synthetic data curation for post-training and structured data extraction