#自然语言处理# A PyTorch implementation of the Transformer model in "Attention is All You Need".
翻译 - “注意就是您所需要的一切”中Transformer模型的PyTorch实现。
A TensorFlow Implementation of the Transformer: Attention Is All You Need
翻译 - 变压器的TensorFlow实现:注意就是您所需要的
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pret...
翻译 - 我对原始变压器模型的实现(Vaswani等)。另外,我还包括了parker.py文件,用于可视化本来似乎很难的概念。当前包括IWSLT预训练模型。
list of efficient attention modules
翻译 - 有效关注模块列表
#大语言模型# Implementation of plug in and play Attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens"