Highlights
- Pro
Stars
Official implementation of "Variable-Rate Deep Image Compression through Spatially-Adaptive Feature Transform", ICCV 2021
A Flexible and Energy Efficient Accelerator For Sparse Convolution Neural Network
[NeurIPS 2024] VFIMamba: Video Frame Interpolation with State Space Models
VMamba: Visual State Space Models,code is based on mamba
Code for paper LocalMamba: Visual State Space Model with Windowed Selective Scan
A PyTorch implementation of the paper "ZigMa: A DiT-Style Mamba-based Diffusion Model" (ECCV 2024)
[BMVC 2024] PlainMamba: Improving Non-hierarchical Mamba in Visual Recognition
Space-filling curve for rectangular domains or arbitrary size.
CVPR2023 - Activating More Pixels in Image Super-Resolution Transformer Arxiv - HAT: Hybrid Attention Transformer for Image Restoration
Implementation of our AAAI'22 work: 'UCTransNet: Rethinking the Skip Connections in U-Net from a Channel-wise Perspective with Transformer'.
[ICML'21 Oral] I-BERT: Integer-only BERT Quantization
Network on Chip Implementation written in SytemVerilog
[CVPR 2023] Official code release of our paper "BiFormer: Vision Transformer with Bi-Level Routing Attention"
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.
A paper list of some recent Transformer-based CV works.
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (V…
Collect some papers about transformer with vision. Awesome Transformer with Computer Vision (CV)
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
Official repository for "Revisiting Weakly Supervised Pre-Training of Visual Perception Models". https://arxiv.org/abs/2201.08371.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Python code for ICLR 2022 spotlight paper EViT: Expediting Vision Transformers via Token Reorganizations