video-representation-learning
There are 19 repositories under video-representation-learning topic.
MCG-NJU/VideoMAE
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
cvlab-columbia/hyperfuture
Code for the paper Learning the Predictability of the Future (CVPR 2021)
xyzforever/BEVT
PyTorch implementation of BEVT (CVPR 2022) https://arxiv.org/abs/2112.01529
ttengwang/Awesome_Long_Form_Video_Understanding
Awesome papers & datasets specifically focused on long-term videos.
ruiwang2021/mvd
[CVPR2023] Masked Video Distillation: Rethinking Masked Feature Modeling for Self-supervised Video Representation Learning (https://arxiv.org/abs/2212.04500)
GV1028/videogan
Implementation of "Generating Videos with Scene Dynamics" in Tensorflow
boschresearch/rince
This is the code accompanying the AAAI 2022 paper "Ranking Info Noise Contrastive Estimation: Boosting Contrastive Learning via Ranked Positives" https://arxiv.org/abs/2201.11736 . The method allows you to use additional ranking information for representation learning.
xiaojieli0903/MaskAgain
Official repository of the “Mask Again: Masked Knowledge Distillation for Masked Video Modeling” (ACM MM 2023)
sunilhoho/VideoMS
Official Pytorch implementation of Efficient Video Representation Learning via Masked Video Modeling with Motion-centric Token Selection.
xiaojieli0903/FGKVMemPred_video
Official repository of the "Fine-grained Key-Value Memory Enhanced Predictor for Video Representation Learning" (ACM MM 2023)
mondalanindya/MSQNet
Actor-agnostic Multi-label Action Recognition with Multi-modal Query [ICCVW '23]
Video-MAC/VideoMAC
Official code for CVPR2024 “VideoMAC: Video Masked Autoencoders Meet ConvNets”
furushchev/chainervr
Chainer implementation of Networks for Learning Video Representations
UARK-AICV/Video_Representation
[Asilomar 2022] Contextual Explainable Video Representation: Human Perception-based Understanding
gimpong/AAAI24-GMMFormer
The code for the paper "GMMFormer: Gaussian-Mixture-Model Based Transformer for Efficient Partially Relevant Video Retrieval" (AAAI'24)
Mallory24/cae_dataset
The official repository for creating casual action effect (CAE) dataset for the IJCNLP-AACL 2023 paper: Implicit Affordance Acquisition via Causal Action–Effect Modeling in the Video Domain
Mallory24/cae_modeling
The official repository for the IJCNLP-AACL 2023 paper: Implicit Affordance Acquisition via Causal Action–Effect Modeling in the Video Domain
XFeiF/ComputerVision_PaperNotes
📚 Paper Notes (Computer vision)
mdnuruzzamanKALLOL/VideoMAE_Tensorflow
VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training