A collection of resources on multimodal learning research.
🐌 Markdown Format:
-
Survey
- (arXiv preprint 2021) A Survey on Multi-modal Summarization, Anubhav Jangra et al. [v1](2021.09.11)
-
2021
- (ICCV 2021 Oral) [Text-guided Image Manipulation] StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery, Or Patashnik et al. [Paper] [Code] [Play]
- (ICCV 2021) [Facial Editing] Talk-to-Edit: Fine-Grained Facial Editing via Dialog, Yuming Jiang et al. [Paper] [Code] [Project] [Dataset Project] [Dataset(CelebA-Dialog Dataset)]
- (arXiv preprint 2021) [Video Action Recognition] ActionCLIP: A New Paradigm for Video Action Recognition, Mengmeng Wang et al. [Paper]
-
2020
-
Yutong ZHOU in Interaction Laboratory, Ritsumeikan University. ଘ(੭*ˊᵕˋ)੭
-
If you have any question, please feel free to contact Yutong ZHOU (E-mail: zhou@i.ci.ritsumei.ac.jp).