guoyanhui03's repositories
ALBEF
Code for ALBEF: a new vision-language pre-training method
AudioCLIP
Source code for models described in the paper "AudioCLIP: Extending CLIP to Image, Text and Audio" (https://arxiv.org/abs/2106.13043)
CLIP
Contrastive Language-Image Pretraining
ConDigSum
Code for EMNLP 2021 paper "Topic-Aware Contrastive Learning for Abstractive Dialogue Summarization"
DALLE-pytorch
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
datasets
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
google-research
Google Research
gupshup
GupShup: Summarizing Open-Domain Code-Switched Conversations EMNLP 2021
lyrebird-wav2clip
Official implementation of the paper WAV2CLIP: LEARNING ROBUST AUDIO REPRESENTATIONS FROM CLIP
natural-language-image-search
Search photos on Unsplash using natural language
pytorch-vq-vae
PyTorch implementation of VQ-VAE by Aäron van den Oord et al.
sonnet
TensorFlow-based neural network library
stylegan2-ada-pytorch
StyleGAN2-ADA - Official PyTorch implementation
TCL
code for TCL: Vision-Language Pre-Training with Triple Contrastive Learning, CVPR 2022
text_summurization_abstractive_methods
Multiple implementations for abstractive text summurization , using google colab
train-CLIP
A PyTorch Lightning solution to training OpenAI's CLIP from scratch.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
vit-pytorch
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
vq-vae-2-pytorch
Implementation of Generating Diverse High-Fidelity Images with VQ-VAE-2 in PyTorch
VQGAN-CLIP
Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.