SOTAVerified

Video Instruction Tuning With Synthetic Data

2024-10-03Unverified0· sign in to hype

Yuanhan Zhang, Jinming Wu, Wei Li, Bo Li, Zejun Ma, Ziwei Liu, Chunyuan Li

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

The development of video large multimodal models (LMMs) has been hindered by the difficulty of curating large amounts of high-quality raw data from the web. To address this, we propose an alternative approach by creating a high-quality synthetic dataset specifically for video instruction-following, namely LLaVA-Video-178K. This dataset includes key tasks such as detailed captioning, open-ended question-answering (QA), and multiple-choice QA. By training on this dataset, in combination with existing visual instruction tuning data, we introduce LLaVA-Video, a new video LMM. Our experiments demonstrate that LLaVA-Video achieves strong performance across various video benchmarks, highlighting the effectiveness of our dataset. We plan to release the dataset, its generation pipeline, and the model checkpoints.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
NExT-QALLaVA-VideoAccuracy83.2Unverified
TVBenchLLaVA-Video 72BAverage Accuracy50Unverified
TVBenchLLaVA-Video 7BAverage Accuracy45.6Unverified

Reproductions