Awesome papers & datasets specifically focused on long-term videos.
-
Updated
Oct 9, 2025
Awesome papers & datasets specifically focused on long-term videos.
Source code for "Bi-modal Transformer for Dense Video Captioning" (BMVC 2020)
End-to-End Dense Video Captioning with Parallel Decoding (ICCV 2021)
[NeurIPS 2023 D&B] VidChapters-7M: Video Chapters at Scale
Official Tensorflow Implementation of the paper "Bidirectional Attentive Fusion with Context Gating for Dense Video Captioning" in CVPR 2018, with code, model and prediction results.
[ICLR 2025] TRACE: Temporal Grounding Video LLM via Casual Event Modeling
PyTorch implementation of Multi-modal Dense Video Captioning (CVPR 2020 Workshops)
[AAAI 2025] VTG-LLM: Integrating Timestamp Knowledge into Video LLMs for Enhanced Video Temporal Grounding
Second-place solution to dense video captioning task in ActivityNet Challenge (CVPR 2020 workshop)
[CVPR 2024] Do you remember? Dense Video Captioning with Cross-Modal Memory Retrieval
Dense video captioning in PyTorch
Official implementation for paper Learning Grounded Vision-Language Representation for Versatile Understanding in Untrimmed Videos
Official Implementation (Pytorch) of the "VidChain: Chain-of-Tasks with Metric-based Direct Preference Optimization for Dense Video Captioning", AAAI 2025
Semantic Metadata Extraction from Generated Video Captions (CD-MAKE 2023).
Add a description, image, and links to the dense-video-captioning topic page so that developers can more easily learn about it.
To associate your repository with the dense-video-captioning topic, visit your repo's landing page and select "manage topics."