#大语言模型#Official release of InternLM series (InternLM, InternLM2, InternLM2.5, InternLM3).
#大语言模型#Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
#大语言模型#[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
Large Context Attention
#计算机科学#Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch
Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch
#大语言模型#LongCite: Enabling LLMs to Generate Fine-grained Citations in Long-context QA
#大语言模型#LLM KV cache compression made easy
#计算机科学#Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
#大语言模型#The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory"
Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
#计算机科学#PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)
#大语言模型#[EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs
✨✨Long-VITA: Scaling Large Multi-modal Models to 1 Million Tokens with Leading Short-Context Accuracy
#大语言模型#[COLM 2024] TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
#大语言模型#awesome llm plaza: daily tracking all sorts of awesome topics of llm, e.g. llm for coding, robotics, reasoning, multimod etc.
open-source code for paper: Retrieval Head Mechanistically Explains Long-Context Factuality
#大语言模型#ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models
#大语言模型#LongQLoRA: Extent Context Length of LLMs Efficiently