要读的论文
后面标记是否读完或者不读了
-
AN IMAGE IS WORTH 16X16 WORDS: TRANSFORMERS FOR IMAGE RECOGNITION AT SCALE
【已完成】笔记:2022.11.18 -
CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification
【已完成】笔记:2022.11.19 -
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
-
ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks
-
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions
-
Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet