Teng Wang's repositories
Caption-Anything
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
Awesome_Prompting_Papers_in_Computer_Vision
A curated list of prompt-based paper in computer vision and vision-language learning.
dense-video-captioning-pytorch
Second-place solution to dense video captioning task in ActivityNet Challenge (CVPR 2020 workshop)
Awesome_Long_Form_Video_Understanding
Awesome papers & datasets specifically focused on long-term videos.
awesome-multimodal-ml
Reading list for research topics in multimodal machine learning
awesome-Vision-and-Language-Pre-training
Recent Advances in Vision and Language Pre-training (VLP)
awesome-vision-language-pretraining-papers
Recent Advances in Vision and Language PreTrained Models (VL-PTMs)
densevid_eval
Evaluation code for Dense-Captioning Events in Videos
ENAS-pytorch
PyTorch implementation of "Efficient Neural Architecture Search via Parameters Sharing"
faster-rcnn.pytorch
A faster pytorch implementation of faster r-cnn
ImageCaptioning.pytorch
image captioning codebase in pytorch(finetunable cnn in branch "with_finetune";diverse beam search can be found in 'dbs' branch; self-critical training is under my self-critical.pytorch repository.)
PrefixTuning
Prefix-Tuning: Optimizing Continuous Prompts for Generation
PromptPapers
Must-read papers on prompt-based tuning for pre-trained language models.
pytorch-CycleGAN-and-pix2pix
Image-to-image translation in PyTorch (e.g. horse2zebra, edges2cats, and more)
self-critical.pytorch
Unofficial pytorch implementation for Self-critical Sequence Training for Image Captioning. and others.
slowfast_feature_extractor
Feature Extractor module for videos using the PySlowFast framework