qpmnh's repositories
acoustic-images-self-supervision
Code for the paper "Leveraging Acoustic Images for Effective Self-Supervised Audio Representation Learning" ECCV 2020
solo-learn
solo-learn: a library of self-supervised methods for visual representation learning powered by Pytorch Lightning
-Audio-visualization
使用快速傅里叶变换(FFT)实现的音频文件的可视化
AVVP-ECCV20
Unified Multisensory Perception: Weakly-Supervised Audio-Visual Video Parsing, ECCV, 2020. (Spotlight)
awesome-audio-visual
A curated list of different papers and datasets in various areas of audio-visual processing
awesome-computer-vision
A curated list of awesome computer vision resources
awesome-self-supervised-learning
A curated list of awesome self-supervised methods
CM-ACC
Cross-model active contrastive coding
deep_avsr
A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.
DeepClustering
Methods and Implements of Deep Clustering
KWS-Net
Seeing Wake Words: Audio-visual Keyword Spotting
Localizing-Visual-Sounds-the-Hard-Way
Localizing Visual Sounds the Hard Way
Mead
MEAD: A Large-scale Audio-visual Dataset for Emotional Talking-face Generation [ECCV2020]
moco
PyTorch implementation of MoCo: https://arxiv.org/abs/1911.05722
OpenSelfSup
Self-Supervised Learning Toolbox and Benchmark
PSOL
code repository of “Rethinking the Route Towards Weakly Supervised Object Localization” in CVPR 2020
PSP_CVPR_2021
A PyTorch implementation of the CVPR-2021 paper: Positive Sample Propagation along the Audio-Visual Event Line.
qpmnh.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
Separating-Sounds-from-a-Single-Image
PyTorch implementation of "Separating Sounds from a Single Image"
sound-spaces
A first-of-its-kind acoustic simulation platform for audio-visual embodied AI research. It supports training and evaluating multiple tasks and applications.
Talking-Face_PC-AVS
Code for Pose-Controllable Talking Face Generation by Implicitly Modularized Audio-Visual Representation (CVPR 2021)
VisualVoice
Audio-Visual Speech Separation with Cross-Modal Consistency