Pinned Repositories
-Audio-visualization
使用快速傅里叶变换(FFT)实现的音频文件的可视化
111
2.5D-Visual-Sound
2.5D visual sound
3D-ResNets-PyTorch
3D ResNets for Action Recognition (CVPR 2018)
academic-resume
acoustic-images-self-supervision
Code for the paper "Leveraging Acoustic Images for Effective Self-Supervised Audio Representation Learning" ECCV 2020
Adaptive
Pytorch Implementation of Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning
AdaptiveAttention
Implementation of "Knowing When to Look: Adaptive Attention via A Visual Sentinel for Image Captioning"
mvda
Discriminant Analysis Algorithms
solo-learn
solo-learn: a library of self-supervised methods for visual representation learning powered by Pytorch Lightning
qpmnh's Repositories
qpmnh/acoustic-images-self-supervision
Code for the paper "Leveraging Acoustic Images for Effective Self-Supervised Audio Representation Learning" ECCV 2020
qpmnh/solo-learn
solo-learn: a library of self-supervised methods for visual representation learning powered by Pytorch Lightning
qpmnh/-Audio-visualization
使用快速傅里叶变换(FFT)实现的音频文件的可视化
qpmnh/111
qpmnh/academic-resume
qpmnh/AudioDVP
AudioDVP:Photorealistic Audio-driven Video Portraits
qpmnh/AVID-CMA
Audio Visual Instance Discrimination with Cross-Modal Agreement
qpmnh/AVVP-ECCV20
Unified Multisensory Perception: Weakly-Supervised Audio-Visual Video Parsing, ECCV, 2020. (Spotlight)
qpmnh/awesome-audio-visual
A curated list of different papers and datasets in various areas of audio-visual processing
qpmnh/awesome-computer-vision
A curated list of awesome computer vision resources
qpmnh/awesome-self-supervised-learning
A curated list of awesome self-supervised methods
qpmnh/CM-ACC
Cross-model active contrastive coding
qpmnh/deep_avsr
A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.
qpmnh/DeepClustering
Methods and Implements of Deep Clustering
qpmnh/KWS-Net
Seeing Wake Words: Audio-visual Keyword Spotting
qpmnh/Localizing-Visual-Sounds-the-Hard-Way
Localizing Visual Sounds the Hard Way
qpmnh/Mead
MEAD: A Large-scale Audio-visual Dataset for Emotional Talking-face Generation [ECCV2020]
qpmnh/moco
PyTorch implementation of MoCo: https://arxiv.org/abs/1911.05722
qpmnh/OpenSelfSup
Self-Supervised Learning Toolbox and Benchmark
qpmnh/PSOL
code repository of “Rethinking the Route Towards Weakly Supervised Object Localization” in CVPR 2020
qpmnh/PSP_CVPR_2021
A PyTorch implementation of the CVPR-2021 paper: Positive Sample Propagation along the Audio-Visual Event Line.
qpmnh/qpmnh.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
qpmnh/Separating-Sounds-from-a-Single-Image
PyTorch implementation of "Separating Sounds from a Single Image"
qpmnh/sound-spaces
A first-of-its-kind acoustic simulation platform for audio-visual embodied AI research. It supports training and evaluating multiple tasks and applications.
qpmnh/starter-hugo-academic
qpmnh/starter-hugo-portfolio-theme
qpmnh/Talking-Face_PC-AVS
Code for Pose-Controllable Talking Face Generation by Implicitly Modularized Audio-Visual Representation (CVPR 2021)
qpmnh/VGGSound
VGGSound: A Large-scale Audio-Visual Dataset
qpmnh/VisualVoice
Audio-Visual Speech Separation with Cross-Modal Consistency
qpmnh/WAL