Transform Video as a Document with ChatGPT, CLIP, BLIP2, GRIT, Whisper, LangChain.
-
Updated
Jul 25, 2023 - Python
Transform Video as a Document with ChatGPT, CLIP, BLIP2, GRIT, Whisper, LangChain.
An official implementation for " UniVL: A Unified Video and Language Pre-Training Model for Multimodal Understanding and Generation"
[CVPR2022] Official Implementation of ReferFormer
[ICCV2023] UniVTG: Towards Unified Video-Language Temporal Grounding
[CVPR2023] All in One: Exploring Unified Video-Language Pre-training
[NeurIPS2022] Egocentric Video-Language Pretraining
The repository collects many various multi-modal transformer architectures, including image transformer, video transformer, image-language transformer, video-language transformer and self-supervised learning models. Additionally, it also collects many useful tutorials and tools in these related domains.
Align and Prompt: Video-and-Language Pre-training with Entity Prompts
Pytorch code for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
A new multi-shot video understanding benchmark Shot2Story with comprehensive video summaries and detailed shot-level captions.
[CVPR21] Visual Semantic Role Labeling for Video Understanding (https://arxiv.org/abs/2104.00990)
A Survey on video and language understanding.
The Pytorch implementation for "Video-Text Pre-training with Learned Regions"
PyTorch code for "Perceiver-VL: Efficient Vision-and-Language Modeling with Iterative Latent Attention" (WACV 2023)
A curated list of video-text datasets in a variety of languages. These datasets can be used for video captioning (video description) or video retrieval.
Official implementation for paper Learning Grounded Vision-Language Representation for Versatile Understanding in Untrimmed Videos
[EMNLP 2024] A Video Chat Agent with Temporal Prior
[NeurIPS2024] VideoGUI: A Benchmark for GUI Automation from Instructional Videos
Code for CVPR 2023 paper "SViTT: Temporal Learning of Sparse Video-Text Transformers"
[ICCV 2023] The official PyTorch implementation of the paper: "Localizing Moments in Long Video Via Multimodal Guidance"
Add a description, image, and links to the video-language topic page so that developers can more easily learn about it.
To associate your repository with the video-language topic, visit your repo's landing page and select "manage topics."