qhfan's repositories
segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
argoverse-api
Official GitHub repository for Argoverse dataset
BiFormer
[CVPR 2023] Official code release of our paper "BiFormer: Vision Transformer with Bi-Level Routing Attention"
bleurt
BLEURT is a metric for Natural Language Generation based on transfer learning.
deit
Official DeiT repository
EdgeNeXt
[CADL'22, ECCVW] Official repository of paper titled "EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for Mobile Vision Applications".
Efficient-AI-Backbones
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
FocalNet
[NeurIPS 2022] Official code for "Focal Modulation Networks"
fourier_neural_operator
Use Fourier transform to learn operators in differential equations.
FSQ-pytorch
A Pytorch Implementation of Finite Scalar Quantization
insightface
State-of-the-art 2D and 3D Face Analysis Project
LITv2
[NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "Fast Vision Transformers with HiLo Attention"
MIRL
[NeurIPS 2023] Masked Image Residual Learning for Scaling Deeper Vision Transformers
ml-cvnets
CVNets: A library for training computer vision networks
Neighborhood-Attention-Transformer
Official NAT and DiNAT repository.
open_clip
An open source implementation of CLIP.
openmixup
CAIRI Supervised, Semi- and Self-Supervised Visual Representation Learning Toolbox and Benchmark
pytorch-image-models
PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, EfficientNetV2, NFNet, Vision Transformer, MixNet, MobileNet-V3/V2, RegNet, DPN, CSPNet, and more
U-ViT
A PyTorch implementation of the paper "All are Worth Words: A ViT Backbone for Diffusion Models".
ViTAE-VSA
This is an official implementation for "VSA: Learning Varied-Size Window Attention in Vision Transformers"