video-captioning
There are 86 repositories under video-captioning topic.
YehLi/xmodaler
X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsense reasoning, and cross-modal retrieval).
xiadingZ/video-caption.pytorch
pytorch implementation of video captioning
scopeInfinity/Video2Description
Video to Text: Natural language description generator for some given video. [Video Captioning]
tomchang25/whisper-auto-transcribe
Auto transcribe tool based on whisper
antoyang/VidChapters
[NeurIPS 2023 D&B] VidChapters-7M: Video Chapters at Scale
jayleicn/recurrent-transformer
[ACL 2020] PyTorch code for MART: Memory-Augmented Recurrent Transformer for Coherent Video Paragraph Captioning
vijayvee/video-captioning
This repository contains the code for a video captioning system inspired by Sequence to Sequence -- Video to Text. This system takes as input a video and generates a caption in English describing the video.
JasonYao81000/MLDS2018SPRING
Machine Learning and having it Deep and Structured (MLDS) in 2018 spring
jpthu17/EMCL
[NeurIPS 2022 Spotlight] Expectation-Maximization Contrastive Learning for Compact Video-and-Language Representations
jssprz/video_captioning_datasets
Summary about Video-to-Text datasets. This repository is part of the review paper *Bridging Vision and Language from the Video-to-Text Perspective: A Comprehensive Review*
terry-r123/Awesome-Captioning
A curated list of Multimodal Captioning related research(including image captioning, video captioning, and text captioning)
bytedance/Shot2Story
A new multi-shot video understanding benchmark Shot2Story with comprehensive video summaries and detailed shot-level captions.
jayleicn/TVCaption
[ECCV 2020] PyTorch code of MMT (a multimodal transformer captioning model) on TVCaption dataset
Kamino666/Video-Captioning-Transformer
这是一个基于Pytorch平台、Transformer框架实现的视频描述生成 (Video Captioning) 深度学习模型。 视频描述生成任务指的是:输入一个视频,输出一句描述整个视频内容的文字(前提是视频较短且可以用一句话来描述)。本repo主要目的是帮助视力障碍者欣赏网络视频、感知周围环境,促进“无障碍视频”的发展。
nasib-ullah/video-captioning-models-in-Pytorch
A PyTorch implementation of state of the art video captioning models from 2015-2019 on MSVD and MSRVTT datasets.
UARK-AICV/VLTinT
[AAAI 2023 Oral] VLTinT: Visual-Linguistic Transformer-in-Transformer for Coherent Video Paragraph Captioning
ParitoshParmar/MTL-AQA
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
amazon-science/crossmodal-contrastive-learning
CrossCLR: Cross-modal Contrastive Learning For Multi-modal Video Representations, ICCV 2021
jacobswan1/Video2Commonsense
Video captioning baseline models on Video2Commonsense Dataset.
lvapeab/ABiViRNet
Attention Bidirectional Video Recurrent Net
imshaikot/srt-webvtt
Convert SRT formatted subtitle to WebVTT on the fly over HTML5/browser environment
pochih/Video-Cap
🎬 Video Captioning: ICCV '15 paper implementation
LuoweiZhou/densecap
Dense video captioning in PyTorch
tsujuifu/pytorch_empirical-mvm
A PyTorch implementation of EmpiricalMVM
TXH-mercury/COSA
[ICLR2024] Codes and Models for COSA: Concatenated Sample Pretrained Vision-Language Foundation Model
WingsBrokenAngel/delving-deeper-into-the-decoder-for-video-captioning
Source code for Delving Deeper into the Decoder for Video Captioning
acherstyx/CoCap
[ICCV 2023] Accurate and Fast Compressed Video Captioning
xiadingZ/video-caption-openNMT.pytorch
implement video caption based on openNMT
willyfh/awesome-video-text-datasets
A curated list of video-text datasets in a variety of languages. These datasets can be used for video captioning (video description) or video retrieval.
mlvlab/MELTR
MELTR: Meta Loss Transformer for Learning to Fine-tune Video Foundation Models (CVPR 2023)
jssprz/visual_syntactic_embedding_video_captioning
Source code of the paper titled *Improving Video Captioning with Temporal Composition of a Visual-Syntactic Embedding*
zjr2000/LLMVA-GEBC
Winner solution to Generic Event Boundary Captioning task in LOVEU Challenge (CVPR 2023 workshop)
UARK-AICV/VLCAP
[ICIP 2022] VLCap: Vision-Language with Contrastive Learning for Coherent Video Paragraph Captioning
rohit-gupta/Video2Language
Generating video descriptions using deep learning in Keras
yangbang18/CARE
(TIP'2023) Concept-Aware Video Captioning: Describing Videos with Effective Prior Information
thtang/ADLxMLDS2017
Deep learning works for ADLxMLDS (CSIE 5431) in NTU