Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
翻译 - 在Pytorch中实现视觉变压器,这是仅使用一个变压器编码器即可在视觉分类中实现SOTA的简单方法
#计算机科学#Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
翻译 - 在Pytorch中为CNN和视觉变压器实现了许多类激活图方法。包括Grad-CAM,Grad-CAM ++,Score-CAM,Ablation-CAM和XGrad-CAM
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
翻译 - 这是“Swin Transformer:Hierarchical Vision Transformer using Shifted Windows”的官方实现。
Collect some papers about transformer with vision. Awesome Transformer with Computer Vision (CV)
翻译 - 收集一些有远见的有关变压器的论文。具有计算机视觉 (CV) 功能的出色 Transformer
PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO
翻译 - 使用自学式学习方法DINO进行视觉变形金刚训练的PyTorch代码
Explainability for Vision Transformers
Let's train vision transformers (ViT) for cifar 10!
Vision Transformer (ViT) in PyTorch
Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)
ICCV2021, Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet
EsViT: Efficient self-supervised Vision Transformers
QuadTree Attention for Vision Transformers (ICLR2022)
Awesome Transformers (self-attention) in Computer Vision
Exploring whether attention is necessary for vision transformers
翻译 - 探索视觉变压器是否需要关注
[IEEE TIP] Vision Transformers for Single Image Dehazing
A list of 3D computer vision papers with Transformers
PyTorch Implementation of CvT: Introducing Convolutions to Vision Transformers
#计算机科学#[ICML 2023] Official PyTorch implementation of Global Context Vision Transformers
Polyp-PVT: Polyp Segmentation with Pyramid Vision Transformers, AIR 2023.
(ICLR 2022 Spotlight) Official PyTorch implementation of "How Do Vision Transformers Work?"
Official PyTorch implementation of FasterViT: Fast Vision Transformers with Hierarchical Attention
[NeurIPS 2021] [T-PAMI] DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification
This is an official implementation of CvT: Introducing Convolutions to Vision Transformers.