OpenMMLab Detection Toolbox and Benchmark
-
Updated
Jun 16, 2023 - Python
OpenMMLab Detection Toolbox and Benchmark
This repository contains demos I made with the Transformers library by HuggingFace.
pix2tex: Using a ViT to convert images of equations into LaTeX code.
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
SwinIR: Image Restoration Using Swin Transformer (official repository)
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
OpenMMLab Pre-training Toolbox and Benchmark
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Scenic: A Jax Library for Computer Vision Research and Beyond
An all-in-one toolkit for computer vision
[CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize classifications by Transformer based networks.
EVA Series: Visual Representation Fantasies from BAAI
This is a collection of our NAS and Vision Transformer work.
ICCV2021, Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet
VRT: A Video Restoration Transformer (official repository)
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
The official repo for [NeurIPS'22] "ViTPose: Simple Vision Transformer Baselines for Human Pose Estimation" and [Arxiv'22] "ViTPose+: Vision Transformer Foundation Model for Generic Body Pose Estimation"
You Only Look at One Sequence (NeurIPS 2021)
(ICLR 2022 Spotlight) Official PyTorch implementation of "How Do Vision Transformers Work?"
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions
Add a description, image, and links to the vision-transformer topic page so that developers can more easily learn about it.
To associate your repository with the vision-transformer topic, visit your repo's landing page and select "manage topics."