ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (2024)

Lin Chen1,4 Xilin Wei4∗ Jinsong Li2,4∗ Xiaoyi Dong2,4 Pan Zhang4 Yuhang Zang4
Zehui Chen1,4 Haodong Duan4 Bin Lin3 Zhenyu Tang3
Li Yuan3 Yu Qiao4 Dahua Lin2,4 Feng Zhao1† Jiaqi Wang4†
1 USTC2 CUHK3 PKU4 Shanghai AI Lab
https://sharegpt4video.github.io/

*Equal contribution. Corresponding author.

Abstract

We present the ShareGPT4Video series, aiming to facilitate the video understanding of large video-language models (LVLMs) and the video generation of text-to-video models (T2VMs) via dense and precise captions.The series comprises: 1) ShareGPT4Video, 40K GPT4V annotated dense captions of videos with various lengths and sources, developed through carefully designed data filtering and annotating strategy.2) ShareCaptioner-Video, an efficient and capable captioning model for arbitrary videos, with 4.8M high-quality aesthetic videos annotated by it.3) ShareGPT4Video-8B, a simple yet superb LVLM that reached SOTA performance on three advancing video benchmarks.To achieve this, taking aside the non-scalable costly human annotators, we find using GPT4V to caption video with a naive multi-frame or frame-concatenation input strategy leads to less detailed and sometimes temporal-confused results. We argue the challenge of designing a high-quality video captioning strategy lies in three aspects:1) Inter-frame precise temporal change understanding.2) Intra-frame detailed content description.3) Frame-number scalability for arbitrary-length videos. To this end, we meticulously designed a differential video captioning strategy, which is stable, scalable, and efficient for generating captions for videos with arbitrary resolution, aspect ratios, and length.Based on it, we construct ShareGPT4Video, which contains 40K high-quality videos spanning a wide range of categories, and the resulting captions encompass rich world knowledge, object attributes, camera movements, and crucially, detailed and precise temporal descriptions of events.Based on ShareGPT4Video, we further develop ShareCaptioner-Video, a superior captioner capable of efficiently generating high-quality captions for arbitrary videos. We annotated 4.8M aesthetically appealing videos by it and verified their effectiveness on a 10-second text2video generation task.For video understanding, we verified the effectiveness of ShareGPT4Video on several current LVLM architectures and presented our superb new LVLM ShareGPT4Video-8B. All the models, strategies, and annotations111We do not hold the copyright for any video and will provide the link-annotation pair for research-only usage. will be open-sourced and we hope this project can serve as a pivotal resource for advancing both the LVLMs and T2VMs community.

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (1)

1 Introduction

Recent advancements in multi-modal learning, driven by large language models, have led to progress in image-text dialogue[35, 10, 16, 56, 63, 13, 2, 65] and text-to-image generation tasks[3, 52, 6, 53, 70, 8]. This has inspired a shift towards video understanding[34, 1, 32, 69, 42, 40] and generation tasks[54, 20, 60, 57, 4, 39], allowing for user interaction across video and language modalities.Thus, the detailed and high-fidelity video captions, which bridge the aforementioned modalities, is instrumental in propelling the advancements within the field.

Despite the rich semantic and temporal content of videos, they are often paired with brief captions in existing data. These short descriptions limit the detailed video understanding and the controllability of the video generation. While the importance of detailed captions is recognized in image-text dialogue[10, 7, 55] and text-to-image generation tasks[8, 22], similar efforts are lacking in video understanding and generation.

However, creating large-scale, high-quality video captions is a challenging task. Detailed captioning for long videos is non-trivial and time-consuming even for humans, hindering large-scale annotation. Current open-source LVLMs lack this capability, and closed-source APIs do not yet support video inputs.

On the other hand, if we roughly degrade the input from video to multiple frames, even GPT4V struggles to describe the video with satisfied quality. For example, an intuitive idea is to provide multiple frames with timestamps to the GPT4V and generate the caption, while we find that GPT4V is unstable and sometimes misunderstands the temporal relation between the frames, and its performance further degrades with the increasing of video frame number. Other solutions such as concatenating all the frames into a large image are nonhelpful to the temporal problem, and the caption loses details as the frame number increases. We also showcase these problems in Figure 11-12

We posit that the challenge of devising an effective video captioning strategy is rooted in three fundamental aspects:1) Inter-frame precise temporal change understanding: The temporal dimension distinguishes videos from images. An imprecise temporal description can significantly diminish the quality of the video caption and lead to confusion in the training models.2) Intra-frame detailed content description: Detailed descriptions [10] are crucial for aligning modalities between image and text, which are also important for video-text alignment.3) Frame-number scalability for arbitrary-length videos: Videos encountered in the wild can vary greatly in length. An ideal captioning strategy should be resilient to this variability and generate appropriate captions for videos of any length.

To this end, we present the Differential Sliding-Window Captioning strategy (DiffSW), which is stable, scalable, and efficient for generating captions for arbitrary videos. The central concept of DiffSW is translating the all-frames-to-caption task into a differential description task. Specifically, we generate a detailed caption for the first frame and apply a sliding window of length two to the subsequent frames in chronological order. The powerful image multi-modal model, GPT4V [46], is tasked with identifying the changes between frames based on three inputs: the previous frame, its differential caption, and the current frame. This encompasses alterations in camera movement, object movement, character actions, and scene transitions. Upon acquiring all differential captions, these are input into GPT4 [45] to construct a comprehensive caption for the entire video.The differential concept allows DiffSW to concentrate on the changes between frames, i.e., the temporal changes. Its sliding design ensures the correctness of temporal order and invariance towards the total number of frames. The constant input frame number guarantees that GPT4V does not overlook details and utilizes the API efficiently, resulting in stable, scalable, and efficient caption quality from DiffSW. Furthermore, the differential design enables the re-caption of any sub-clips of a captioned video by reusing its differential captions.

Based on DiffSW, we construct ShareGPT4Video, which contains 40K high-quality video-caption pairs spanning a wide range of categories, and the resulting captions encompass rich world knowledge, object attributes, camera movements, and crucially, detailed and precise temporal descriptions of events.The videos of ShareGPT4Video are collected from various sources [12, 66, 48, 18, 49, 43], employed with a Semantic-based Data Filtering strategy to mitigate content hom*ogeneity among these videos. A Semantic-aware Key-frame Extraction strategy is then applied to the videos to reduce the temporal redundancy. DiffSW is applied to the keyframes to generate high-quality captions and we further improve its stability and quality with a Hierarchical Prompt Design. Manual quality inspection is employed to ensure the quality of the video captions.

Based on ShareGPT4Video, we present ShareCaptionor-Video, an exceptional video captioner capable of efficiently generating high-quality captions for videos of a wide range of resolution, aspect ratio, and duration. It enables the further scaling of high-quality video caption data with minor cost and satisfactory quality, and we generate high-quality captions for 4.8M aesthetically appealing videos (totaling about 3000 hours) by it.

We conduct extensive experiments in video understanding and generation tasks to demonstrate the value of our high-quality video-caption dataset and our superior video captioner.For video generation, a DiT-based [47] text-to-video model trained on the 4.8M video-captions pairs performs well in generating 10-second high-resolution videos and achieving fine-grained control over content generation.For video understanding, ShareGPT4Video brings consistent performance gain of multiple current LVLMs over multiple benchmarks by replacing a small proportion of training data. We further present ShareGPT4Video-8B, a simple yet superb LVLM that reached SOTA performance on three advancing and comprehensive video benchmarks. The model, strategy, and annotations will be publicly available and we hope this project can serve as a pivotal resource for advancing both the LVLMs and T2VMs community.

2 ShareGPT4Video Dataset

This section provides a detailed exposition of how we construct the ShareGPT4Video dataset. We detail the entire process in Figure 2. In Section 2.1, we describe our methods for collecting and filtering video data. In Section 2.2, we explain how we perform efficient sparse sampling of the videos. In Section 2.3, we detail of how we leverage the multimodal image model GPT-4V to generate high-quality captions for the videos.

2.1 Data Collection

Selection of Data Sources. To serve both video understanding and video generation tasks, we consider the aesthetic quality and content complexity of videos during our collection process. We first consider Panda-70M [12], a high-resolution video dataset sourced from YouTube, featuring clips ranging in one minute. This open-domain source covers diverse areas such as wildlife, cooking, sports, news & TV shows, gaming & 3D rendering. It typically includes complex content and transitions, providing a solid foundation for understanding various real-world scenarios. However, the complexity of these contents and transitions presents a significant challenge for the video generation field. To address this, we also source a large volume of aesthetically appealing videos from some user-uploaded video websites [48, 49, 43]. These videos predominantly consist of scenic views and aesthetically pleasing human activities, involving fewer transitions and simpler events.Finally, we supplement our collection with selected videos from Ego4D [18] and BDD100K [66] to fill the gaps in ego-centric human activities and auto-driving scenarios, ensuring our video sources encompass as many real-world scenes as possible.

Semantic-Based Data Filtering. Although our captioning method can support videos of extended lengths, our collection primarily focuses on videos shorter than two minutes due to the trade-off of the duration and amount of videos. We initially filter out videos from our selected data sources longer than two minutes, leaving videos in two minutes as the candidates. We then introduce a semantic-based data filtering strategy to mitigate content hom*ogeneity among these candidates and maintain diversity in the final video dataset. This approach aims to select videos with significant thematic differences from the pool of candidates to compose our final video collection.Specifically, we first use the Panda-Student [12] model to generate a short caption with one sentence for each candidate video, and then maintain a final pool of video candidates. Whenever a new video V is processed, we encode its corresponding short caption S using the Bert-Base-Uncased [15] language model to obtain the CLS token Pn+11×D, which captures high-level semantic expressions. We then calculate the similarity between this CLS token Pn+1 and the CLS tokens {P1,P2,,Pn} of videos already in the final candidate pool. A new video will only be added to the pool if its maximum similarity is below a predefined threshold. We provide the pseudo-code in Figure 14.

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (2)

2.2 Video Processing

Videos are commonly redundant on the temporal dimension, and keyframe sampling is a general idea to represent a video compactly.However, traditional key-frame extraction methods [74, 5] often struggle to ensure semantic coherence, leading to missing key-frames covering crucial changes and transitions. Consequently, we develop a semantic-aware key-frame extraction method that strikes a balance between reducing temporal redundancy and maintaining semantic coherence.

Semantic-aware Key-frame Extraction.We denote VT×H×W×3 as a T frame set sampled from a video with fixed 2-second intervals.We calculate the keyframe set VT×H×W×3 that are sufficiently sparse yet comprehensively cover the evolution of events within the video that T<T. We view the output CLS token of the CLIP-Large image encoder [50] as the global semantics of each frame and remove the adjacent frames that have a high semantic similarity. In practice, we initialize the keyframe set V with the first frame of V. For each frame in V, we calculate its semantic similarity d with the latest keyframe in V. If d is lower than the pre-defined threshold, we view the frame as a keyframe and add it to the V. If not, the frame is skipped as redundant. For completeness, the last frame of V is always added in V. We provide the pseudo-code in Figure 15.

2.3 Captioning Pipeline

As we mentioned in Section 1, we find if we feed all the frames to the GPT4V directly, the GPT4V struggles to stably generate captions with the correct temporal relation between frames, and its performance further worsens with the frame number increasing. On the other hand, if we concatenate all the frames into a large image, the GPT4V loses more details with the increasing frame number, as shown in Figure 11-12.

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (3)

Differential Sliding-window Captioning.To this end, we develop a differential sliding-window captioning pipeline to generate high-quality captions with detailed temporal descriptions for various videos. Specifically, the input fed to the image multi-modal model each time includes the current key-frame and the previous key-frame along with its differential caption. Then, we introduce the Differential Prompt to guide GPT4V in focusing on the changes between the current and previous frames, such as posture, position, camera angle, etc. Additionally, incorporating the differential caption of the previous frame as supplementary context enhances the response quality and reduces hallucinations. This is because the image embedding and textual caption provide explicit and implicit representations of the image, respectively. The differential caption not only adds extra context but also integrates temporal information from two frames ago, further improving the model’s temporal understanding. It’s important to note that for the first key-frame, which lacks a preceding frame, its differential caption is replaced directly with the standard caption. Finally, we input all differential captions along with their corresponding timestamps into GPT4. A specific Summary Prompt is designed to instruct the LLM to generate high-quality video captions with precise temporal dynamics and detailed spatial information. In practice, we use GPT-4-Turbo-04-09 for all the annotations.

In the design of the prompts, we discovered that an explicit Hierarchical Prompt Design significantly aids the GPT4V in comprehending its role, its expected format, and its operational boundaries. This approach contributes to the stabilization of the output’s format and enhances the overall quality of the results. For more details, please refer to Section A.2

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (4)

3 ShareCaptioner-Video

3.1 Model design

We fine-tune the IXC2-4KHD [17] using the collected video caption data, resulting in our ShareCaptioner-Video. For flexible usage, we re-organize the data for the following capabilities:

1. Fast Captioning The model employs an image-grid format for direct video captioning, providing rapid generation speeds that are ideal for short videos. In practice, we concatenate all the keyframes of a video into a vertically elongated image and train the model on a caption task.

2. Sliding Captioning The model supports streaming captioning in a differential sliding-window format, yielding high-quality captions that are suitable for long videos. Similar to the captioning pipeline used in Section 2.3, we take the two adjacent keyframes alongside the previous differential caption as input, and train the model to describe the events occurring between them.

3. Clip Summarizing The model can swiftly summarize any clip from ShareGPT4Video or videos that have undergone the differential sliding-window captioning process, eliminating the need to re-process frames. We use all the differential descriptions as input, and the output is the video caption.

4. Prompt Re-Captioning: The model can rephrase prompts input by users who prefer specific video generation areas, ensuring that T2VMs trained on high-quality video-caption data maintain format alignment during inference with their training. In practice, we use GPT-4 to generate Sora-style prompts for our dense captions, and we train the re-captioning task in reverse, i.e., by using the generated prompt as input and the dense caption as the training target.

In practice, we fine-tune the model end-to-end over one epoch. We follow the default high-resolution strategy, using ‘HD-55’ for fast captioning and ‘HD-25’ for the others. The learning rate is uniform across all model components and warms up from 0 to 2.5×105 within the first 1% of steps. The batch size is set to 1024, and we sample the data uniformly.

3.2 Scaling-up Captions

To validate the effectiveness of our ShareCaptioner-Video in the video captioning task and further support the development of the video generation domain, we utilized it to annotate a large volume of aesthetically appealing videos. Specifically, we meticulously collect and process 4.8 million video clips, totaling approximately 3000 hours, from three sources: MixKit [43], Pexels [48], and Pixabay [49]. Subsequently, we employ the sliding captioning mode of ShareCaptioner-Video to generate high-quality captions for these videos. The total captioning process requires approximately 4000 H100 GPU hours. We provide some statistics on generated captions in Figure 8.

4 Experiments

4.1 Video Understanding

Datasets and Benchmarks.To thoroughly explore the benefits that our high-quality video-caption data bring to LVLMs, we conduct comprehensive evaluations of the model across three multi-modal video benchmarks. VideoBench [44] curates approximately 15,000 QA pairs spanning 10 evaluation dimensions from 13 existing data sources, such as MSVD-QA [62], MSRVTT-QA [62], Activitynet-QA [67], etc. MVBench [30] is designed to challenge LVLMs with video tasks that cannot be effectively resolved by single-frame reliance, featuring 4,000 QA pairs derived from 11 public video benchmarks. TempCompass [38] specifically assesses the nuanced performance of LVLMs across various temporal aspects, such as speed, direction, and attribute changes. It includes 410 videos and 7,540 meticulously collected instructions, emphasizing temporal comprehension and interaction.

ModelVideoBenchMVBenchTempCompassAvg.
VideoLLaVA-7B [34]34.543.050.642.7
VideoLLaVA-7B+Ours35.243.652.743.8
LLaMA-VID-7B [32]36.541.348.142.0
LLaMA-VID-7B+Ours38.243.250.644.0
LLaMA-VID-13B [32]48.343.351.447.7
LLaMA-VID-13B+Ours52.444.253.350.0

CaptionUnlock ViTVideoBenchMVBenchTempCompassAvg.
×37.347.257.247.2
short×36.947.556.146.8
short37.547.956.947.4
detailed×40.750.360.750.6
detailed41.251.261.551.3

Improving current LVLMs with ShareGPT4Video.We validate the effectiveness of the high-quality video-caption data collected in ShareGPT4Video to improve the performance of current LVLMs. For fairness and simplicity, we integrate 28K high-quality video-caption data related to complex scenes (Panda-70M [12], Ego4D [18], and BDD100K [66]) of ShareGPT4Video to replace the captions data in the VideoChatGPT-100K [42] conversation data with an equivalent number.Then we train the VideoLLaVA [34] and LLaMA-VID [32] with their default training settings and hyperparameters. As shown in Table 1, ShareGPT4Video consistently improves the alignment between video and language modalities in different LVLM architectures and scales. Specifically, VideoLLaVA-7B [34] achieves an average performance gain of 1.1 across three comprehensive multi-modal video benchmarks after integrating high-quality captions, while LLaMA-VID-7B and LLaMA-VID-13B achieve an average gain of 2.0 and 2.3, separately. Our high-quality video-caption data is particularly effective in helping LVLMs achieve significant performance improvements on benchmarks that require complex temporal understanding, such as TempCompass [38].

ShareGPT4Video-8B.To obtain our final ShareGPT4Video-8B model, we start with the LLaVA-Next-8B [26] image multi-modal model. Consistent with previous LVLM approaches [34, 42], we uniformly sample 16 frames from each video and arrange these frames into a 4x4 image grid to form the input for both training and inference, following the IG-VLM [24] strategy. For training data, we first collect 153K VQA data from various instructional video-to-text datasets to build our baseline. This collection includes 13K conversational data from VideoChatGPT [42] and 140K question-answer pairs, with 45K data points from CLEVRER [64], 8K from EGO-QA [18], 34K from NextQA [61], and 53K from TGIF-Transition [31]. Then, these VQA data are combined with 28K video-caption data, forming a consolidated training dataset of 181K samples. For more training details, please refer to Section A.1.

As illustrated in Table 3, 4, 5, we present a quantitative comparison between our ShareGPT4Video-8B model supercharged by our ShareGPT4Video dataset with existing state-of-the-art LVLMs. Notably, compared with previous LVLMs, our ShareGPT4Video-8B attains the most superior performance in all three comprehensive benchmarks.Specifically, thanks to the rich temporal information provided by ShareGPT4Video, our ShareGPT4Video-8B model achieves an impressive average accuracy of 61.5% on the TempCompass benchmark. This is an 11.6% increase over the previous best-performing LVLM, VideoLLaVA-7B.Additionally, despite the VideoBench and MVBench benchmarks collecting a diverse range of QA data from various existing video datasets, we achieve solid performance on these benchmarks, surpassing the previous state-of-the-art by an average accuracy of 2.7% and 8.2%.

Abaltion on caption quality and ViT.Based on ShareGPT4Video-8B, we study how the modality alignment is influenced by caption quality and learnable vision encoder. As indicated in Table 2, introducing short captions on top of VQA data may not yield substantial performance gains. It could even degrade performance on some benchmarks due to sub-optimal modality alignment. Comparing the first, second, and fourth rows of Table 2, the significant performance gains of comprehending temporal sequences benefited from our high-quality caption data are evident. Moreover, unlocking the vision encoder when training with detailed captions facilitates better LVLMs modality alignment.

ModelMulti-Choice QAYes/No QACaption MatchingCaption GenerationAvg.
ACDISPEVATACDISPEVATACDISPEVATACDISPEVAT
Valley-7B [40]47.029.332.518.929.958.152.052.550.352.965.053.852.653.053.854.031.032.734.241.433.4
PandaGPT-13B [12]35.527.829.331.830.953.049.650.853.752.256.651.444.355.049.023.725.726.029.832.640.4
VideoLLaMA-13B [69]54.124.528.132.828.568.146.048.851.850.973.147.447.152.048.354.321.313.938.533.943.3
VideoChatGPT-7B [42]47.031.628.437.130.952.550.049.551.050.064.648.647.849.348.640.928.424.531.833.942.4
mPLUG-Owl-7B [63]66.629.332.234.835.464.450.651.251.352.056.945.346.449.349.046.528.230.431.236.544.5
VideoLLaVA-7B [30]70.432.238.241.439.974.351.850.349.251.188.253.861.957.058.350.828.723.238.233.649.9
LLaMA-VID-7B [32]58.629.929.330.526.063.048.849.248.452.772.745.652.249.049.053.028.021.935.535.944.2
ShareGPT4Video-8B87.634.647.562.964.275.253.858.666.565.693.358.158.875.075.379.832.636.650.853.461.5

ModelANetMSVDMSRVTTTGIFYC2UCFMOTTVMVNBALEDMSQA3DAvg.
mPLUG-Owl-7B [63]41.542.536.331.727.122.827.824.030.225.133.351.032.033.2
Otter-7B [27]44.355.047.034.332.722.416.727.737.134.352.848.729.737.5
Video-LLaMA-7B [69]39.941.234.131.328.927.616.724.832.426.260.649.131.232.8
Valley-7B [40]38.132.028.031.429.120.311.123.732.631.341.756.533.334.0
VideoChat-7B [29]44.642.237.433.727.722.427.826.234.128.639.955.431.435.4
PandaGPT-7B [12]45.050.444.629.733.033.016.727.937.131.141.756.030.837.5
VideoChatGPT-7B [42]46.657.546.335.634.824.127.828.836.522.541.758.237.238.5
ChatUniVi-7B [23]49.048.641.741.329.028.316.723.133.625.738.953.129.135.3
VideoLLaVA-7B* [34]44.134.530.039.430.719.522.227.333.425.633.350.738.934.5
LLaMA-VID-7B* [32]45.244.539.129.129.327.911.134.132.528.936.147.836.836.5
ShareGPT4Video-8B50.845.643.042.834.639.722.231.934.030.541.753.642.941.2

ModelASAPAAFAUAOEOIOSMDALSTACMCMASCFPCOENERCIAvg.
Otter-7B [27]23.023.027.527.029.553.028.033.024.523.527.526.028.518.038.522.022.023.519.019.526.8
mPLUG-Owl-7B [63]22.028.034.029.029.040.527.031.527.023.029.031.527.040.044.024.031.026.020.529.529.7
LLaMA-Adapter [71]23.028.051.030.033.053.532.533.525.521.530.529.022.541.539.525.031.522.528.032.031.7
VideoChatGPT-7B [42]23.526.062.022.526.554.028.040.023.020.031.030.525.539.548.529.033.029.526.035.532.7
VideoLLaMA-7B [69]27.525.551.029.039.048.040.538.022.522.543.034.022.532.545.532.540.030.021.037.034.1
VideoChat-7B [29]33.526.556.033.540.553.040.530.025.527.048.535.020.542.546.026.541.023.523.536.035.5
VideoLLaVA-7B* [34]46.042.556.539.053.553.048.041.029.031.582.545.026.053.041.533.541.527.538.531.543.0
LLaMA-VID-7B* [32]45.540.558.039.555.053.540.035.518.527.587.041.523.045.541.027.040.034.541.531.541.3
ShareGPT4Video-8B49.539.579.540.054.582.554.532.550.541.584.535.562.575.051.025.546.528.539.051.551.2

4.2 Video Captioning

To verify the capability of ShareCapitoner-Video, we quantitatively compare the video captioning quality between ShareCapitoner-Video and GPT4V with human preference voting. As shown in Table 7, it performs on par with GPT4V. We also shows the qualitative results in Figure 9. For more details, please refer to Section A.4

4.3 Video Generation

Model setup.To validate the effectiveness of high-quality captions in the T2VMs area, we utilize ShareCaptioner-Video and Panda-Student [12] to generate high-quality and short video captions for 4.5M videos with 65 frames and 0.3M videos with 221 frames, separately. Following the process outlined in the Open-Sora-Plan [25], we fine-tuned the pretrained T2VM to enable the generation of high-fidelity 10-second videos. For comparison, we fine-tuned a baseline model with the same quantity of video-short-captions pairs. For more training details, please refer to Section A.1.

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (5)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (6)

Qualitative analysis.As illustrated in Figure 5, the T2VM can accurately follow detailed prompts and demonstrate remarkable control over semantic content and camera movement when aided by high-quality, detailed captions generated by ShareCaptioner-Video. The resulting video showcases intricate and lively content. In contrast, when provided with brief captions, the T2VM struggles to adhere to complex generation prompts, leading to subpar results.

5 Limitations and Social Impacts

Limitations. Although our current pipeline for generating high-quality video captions fully utilizes visual and textual information, it is limited by GPT4V’s inability to incorporate audio information simultaneously. Audio information is beneficial in conversational scenarios involving daily human activities. We plan to introduce audio information in future work, once GPT4o supports audio input, to enhance the quality of our captions further.

Social impacts. 1)Since the large language model involves the generation process of the large-scale captions, we have not manually verified each caption for socially biased content; 2) Although we utilize video data from existing public datasets, we cannot ensure that the selected videos do not contain human faces. Therefore, while there are no restrictions on the use of our generated captions, users must adhere to the licenses of the original video sources when using the videos.

6 Conclusion

In this study, we aim to address the challenge of lacking high-quality video-caption data for large video-language models (LVLMs) and text-to-video models (T2VMs). We develop ShareGPT4Video, a high-quality video-caption dataset, and ShareCaptioner-Video, an advanced and versatile model in the video-language multi-modal area. By employing a series of strategies and designs, we generate 40K detailed captions from advanced image multi-modal model, GPT4V, and 4.8M high-quality captions from our ShareCaptioner-Video. These captions include rich world knowledge, object attributes, camera movements, and detailed temporal descriptions of events. Our extensive experiments validate the effectiveness of our dataset and captioner in enhancing video understanding and generation tasks. We believe that ShareGPT4Video and ShareCaptioner-Video will serve as essential resources for advancing research in the LVLM and T2VM communities.

Appendix A Appendix

In the appendix, we provide more results and analysis and summarize them as follows:

  • In Section A.1, we detail the experimental setups.

  • In Section A.2, we talk about the prompt design philosophy and showcase the prompt for each stage.

  • In Section A.3, we present the template of our hierarchical prompt.

  • In Section A.4, we introduce the qualitative and quantitative comparison of the caption quality between our ShareCaptioner-Video and other methods.

  • In Section A.5, we present more statics of the ShareGPT4Video dataset.

  • In Section A.6, we provide the detailed pseudo code of semantic-based data filtering and semantic-aware key-frame extraction.

  • In Section A.7, we present the section of related work.

A.1 Experimental Details

Training details of ShareGPT4Video-8B.During training, we employ a batch size of 128 and the AdamW optimizer. We opt to fine-tune the entire model, setting the learning rate for the vision encoder at 2e-6, for the MLP projector at 2e-5, and for the LLM using LoRA [21], the learning rate is set at 2e-4. Such training strategy enables us to obtain an exceptional LVLM, ShareGPT4Video-8B, with 8 A100 GPUs in about 5 hours.

Implementation details for T2VMs. We utilize the Latte-XL [41] model with pre-trained weights and a text encoder from T5-XXL [51]. In the first stage, we perform pretraining on a lower number of frames and enable joint image-video training, with the image batch size being four times that of the video. This stage involved 50k training steps. In the second stage, both the video and image batch sizes were reduced to 2. For all training stages, we used AdamW optimizer with a constant learning rate of 2e-5, and the resolution for both images and videos was set to 512×512. For training precision, we used Bf16 since fp16 led to loss becoming NaN. The first stage requires around 6K H100 GPU hours, and the second stage requires around 36K Ascend GPU hours.

A.2 Hierarchical Prompt Design.

We introduce this design to help the multi-modal and language models effectively perform their roles during the captioning process. We separately illustrate the differential-caption prompt and summary prompt in Figure 7 and Figure 8.Hierarchical prompts primarily consist of four components. The Character part provides the model with an overall perception of the role it is to play and the work environment it faces. The Skills section specifies the skills the model needs to possess, ensuring precise compliance with multiple requirements without omissions or confusion. The Constraints section clarifies behaviors that users do not desire and the rules to be followed when constructing outputs. The Structured Input section requires users to set up according to their specific scenarios. For example, when guiding the model to generate differential captions, the Character part informs the model that it is an expert in analyzing video frames. The Skills section requires the model to describe inter-frame target actions and behaviors, changes in environment and background, alterations in target appearance attributes, and camera movements reflecting temporal changes. The Constraints section demands precise descriptions without listing them item by item. In the Structured Input section, the input consists of frame indexes, timestamps, the previous frame, and its differential caption, among others.

A.3 Prompt Template

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (7)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (8)

A.4 Comparison of Caption Quality

Lexical Categoryn.adj.adv.v.num.prep.
GPT4V27.5%11.2%2.0%12.3%0.3%11.3%
Share-Captioner28.1%2.8%1.5%12.2%0.5%11.4%

GPT4VShareCaptioner-VideoComparable
Percentage37.8%36.5%23.7%
Avg. Score2.22.0-

Quantitative captioning capability comparison between ShareCaptioner-Video and GPT4V. We first analyze the linguistic composition of the captions produced by GPT4V and ShareCaptioner-Video, and the results are presented in Table 7. The analysis reveals that the captions generated by our ShareCaptioner-Video contain a comparable information level to those generated by GPT4V. Furthermore, as shown in Table 7, we first generate 100 captions with GPT4V and ShareCaptioner-Video and then invite 10 volunteers to evaluate the captions based on three aspects. These aspects include (1) Omission and Fabrication - checking for no key elements missing in the caption and identifying imaged elements not present in the video; (2) Distortion - assessing the accuracy of element attributes such as color and size; and (3) Temporal Mismatch - evaluating whether the description accurately reflects the evolution of temporal events in the video. Each pair can earn a maximum of 3 points, one for each criterion met. As anticipated, our ShareCaptioner-Video performs on par with the GPT4V.

Qualitative comparison of caption quality from various sources. In Figure 9, we compare the quality of captions from different sources. The result which uses the Panda-Student [12] model learned from multiple LVLMs, generates one-sentence captions for videos. While this approach minimizes errors, it provides insufficient textual information for information-dense videos, hindering the alignment of video and language modalities. The Video ChatCaptioner [9], on the other hand, uses ChatGPT to construct questions, employs BLIP2 [28] for VQA, and then uses ChatGPT again to summarize a complete caption based on the QA. Although this method produces longer captions than Panda-Student, the extensive involvement of ChatGPT often leads to over-imagination and hallucinations, negatively impacting the alignment between video and language modalities. Thanks to our carefully designed pipeline for generating high-quality video captions, GPT4V demonstrated impressive captioning capabilities. Additionally, after fine-tuning on the ShareGPT4Video dataset, our ShareCaptioner-Video also exhibits captioning capabilities comparable to GPT-4V.

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (9)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (10)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (11)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (12)

A.5 Data Statistics

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (13)
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (14)

A.6 Pseudo Code

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (15)

ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (16)

A.7 Related Work

Large video-language models. The vision-language community has recently made significant advancements in the field of large image-language models [36, 35, 37, 10, 13, 11, 58, 59, 2, 72, 33]. These models typically employ three core components: a visual encoder for extracting visual features, a connector to bridge the visual and language modalities, and a large language model (LLM) to decode the multimodal context. A milestone in this field is LLaVA-1.5 [35], which uses CLIP-Large [50] as the visual encoder, a two-layer MLP as the connector, and Vicuna-v1.5 [14] as the LLM backbone. This model achieves exceptional performance with only affordable data for training. Following this paradigm, the large video-language model field has also seen emerging efforts to encode video into vision tokens suitable for LLMs, enabling understanding and reasoning about video content. For example, VideoChat2 [30] employs a video transformer and a Q-Former [28] to extract and compress visual tokens. VideoLLaVA [34] uses a pre-aligned video encoder with image and language, along with an MLP, to extract and transform visual tokens. LLaMA-VID [32] introduces the LLaVA-like three-stage training strategy and compresses each frame’s vision token into one to support long video understanding. Despite the continuous exploration of training strategies and model architectures, none of these methods have investigated the role of high-quality captions in aligning video and language modalities, as has been done in the large image-language model field [10, 7].

Text-to-video models. Recent advancements in the text-to-image field, such as DALL·E [3], and Stable Diffusion [52], have significantly propelled the T2I area. Inspired by these achievements, researchers have increasingly begun to explore the potential of generating high-fidelity videos from textual descriptions. For instance, MagicVideo [73] introduces a 3D U-Net-based architecture and directs temporal attention to ensure efficient, high-fidelity video generation while maintaining temporal consistency and authenticity. PixelDance [68] and SVD [4] leverage pixel-based and latent-based techniques to produce high-resolution videos, while Make-A-Video [54] and Imagen Video [19] extend text-to-image models and techniques to the text-to-video domain. A milestone in this field is Sora [39], which ambitiously employs DiT to train a text-to-video model from scratch. Sora is the first model capable of generating minute-long videos based on user instructions. The success of this remarkable feat relies on extensive high-quality video-caption data, meticulously designed architectural optimizations, and substantial computational power. Despite the pivotal role of high-quality video-caption data in T2VMs, this aspect has not received sufficient attention. Therefore, in this work, we aim to construct a high-quality video-caption dataset to advance the development of T2VMs.

References

  • [1]K. Ataallah, X. Shen, E. Abdelrahman, E. Sleiman, D. Zhu, J. Ding, and M. Elhoseiny.Minigpt4-video: Advancing multimodal llms for video understanding with interleaved visual-textual tokens.arXiv preprint arXiv:2404.03413, 2024.
  • [2]J. Bai, S. Bai, S. Yang, S. Wang, S. Tan, P. Wang, J. Lin, C. Zhou, and J. Zhou.Qwen-vl: A frontier large vision-language model with versatile abilities.arXiv preprint arXiv:2308.12966, 2023.
  • [3]J. Betker, G. Goh, L. Jing, T. Brooks, J. Wang, L. Li, L. Ouyang, J. Zhuang, J. Lee, Y. Guo, et al.Improving image generation with better captions.Computer Science. https://cdn. openai. com/papers/dall-e-3. pdf, 2(3):8, 2023.
  • [4]A. Blattmann, T. Dockhorn, S. Kulal, D. Mendelevitch, M. Kilian, D. Lorenz, Y. Levi, Z. English, V. Voleti, A. Letts, et al.Stable video diffusion: Scaling latent video diffusion models to large datasets.arXiv preprint arXiv:2311.15127, 2023.
  • [5]J. Calic and E. Izuierdo.Efficient key-frame extraction and video analysis.In Proceedings. International Conference on Information Technology: Coding and Computing, pages 28–33. IEEE, 2002.
  • [6]H. Chang, H. Zhang, J. Barber, A. Maschinot, J. Lezama, L. Jiang, M.-H. Yang, K. Murphy, W. T. Freeman, M. Rubinstein, et al.Muse: Text-to-image generation via masked generative transformers.arXiv preprint arXiv:2301.00704, 2023.
  • [7]G. H. Chen, S. Chen, R. Zhang, J. Chen, X. Wu, Z. Zhang, Z. Chen, J. Li, X. Wan, and B. Wang.Allava: Harnessing gpt4v-synthesized data for a lite vision-language model.arXiv preprint arXiv:2402.11684, 2024.
  • [8]J. Chen, J. Yu, C. Ge, L. Yao, E. Xie, Y. Wu, Z. Wang, J. Kwok, P. Luo, H. Lu, et al.Pixart-alpha: Fast training of diffusion transformer for photorealistic text-to-image synthesis.arXiv preprint arXiv:2310.00426, 2023.
  • [9]J. Chen, D. Zhu, K. Haydarov, X. Li, and M. Elhoseiny.Video chatcaptioner: Towards the enriched spatiotemporal descriptions.arXiv preprint arXiv:2304.04227, 2023.
  • [10]L. Chen, J. Li, X. Dong, P. Zhang, C. He, J. Wang, F. Zhao, and D. Lin.Sharegpt4v: Improving large multi-modal models with better captions.arXiv preprint arXiv:2311.12793, 2023.
  • [11]L. Chen, J. Li, X. Dong, P. Zhang, Y. Zang, Z. Chen, H. Duan, J. Wang, Y. Qiao, D. Lin, et al.Are we on the right way for evaluating large vision-language models?arXiv preprint arXiv:2403.20330, 2024.
  • [12]T.-S. Chen, A. Siarohin, W. Menapace, E. Deyneka, H.-w. Chao, B. E. Jeon, Y. Fang, H.-Y. Lee, J. Ren, M.-H. Yang, et al.Panda-70m: Captioning 70m videos with multiple cross-modality teachers.arXiv preprint arXiv:2402.19479, 2024.
  • [13]Z. Chen, J. Wu, W. Wang, W. Su, G. Chen, S. Xing, Z. Muyan, Q. Zhang, X. Zhu, L. Lu, et al.Internvl: Scaling up vision foundation models and aligning for generic visual-linguistic tasks.arXiv preprint arXiv:2312.14238, 2023.
  • [14]W.-L. Chiang, Z. Li, Z. Lin, Y. Sheng, Z. Wu, H. Zhang, L. Zheng, S. Zhuang, Y. Zhuang, J. E. Gonzalez, et al.Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality.See https://vicuna. lmsys. org (accessed 14 April 2023), 2023.
  • [15]J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova.Bert: Pre-training of deep bidirectional transformers for language understanding.arXiv preprint arXiv:1810.04805, 2018.
  • [16]X. Dong, P. Zhang, Y. Zang, Y. Cao, B. Wang, L. Ouyang, X. Wei, S. Zhang, H. Duan, M. Cao, et al.Internlm-xcomposer2: Mastering free-form text-image composition and comprehension in vision-language large model.arXiv preprint arXiv:2401.16420, 2024.
  • [17]X. Dong, P. Zhang, Y. Zang, Y. Cao, B. Wang, L. Ouyang, S. Zhang, H. Duan, W. Zhang, Y. Li, et al.Internlm-xcomposer2-4khd: A pioneering large vision-language model handling resolutions from 336 pixels to 4k hd.arXiv preprint arXiv:2404.06512, 2024.
  • [18]K. Grauman, A. Westbury, E. Byrne, Z. Chavis, A. Furnari, R. Girdhar, J. Hamburger, H. Jiang, M. Liu, X. Liu, et al.Ego4d: Around the world in 3,000 hours of egocentric video.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18995–19012, 2022.
  • [19]J. Ho, W. Chan, C. Saharia, J. Whang, R. Gao, A. Gritsenko, D. P. Kingma, B. Poole, M. Norouzi, D. J. Fleet, et al.Imagen video: High definition video generation with diffusion models.arXiv preprint arXiv:2210.02303, 2022.
  • [20]W. Hong, M. Ding, W. Zheng, X. Liu, and J. Tang.Cogvideo: Large-scale pretraining for text-to-video generation via transformers.arXiv preprint arXiv:2205.15868, 2022.
  • [21]E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, and W. Chen.Lora: Low-rank adaptation of large language models.arXiv preprint arXiv:2106.09685, 2021.
  • [22]B. James, G. Gabriel, J. Li, and B. Tim.Improving image generation with better captions.2023.
  • [23]P. Jin, R. Takanobu, C. Zhang, X. Cao, and L. Yuan.Chat-univi: Unified visual representation empowers large language models with image and video understanding.arXiv preprint arXiv:2311.08046, 2023.
  • [24]W. Kim, C. Choi, W. Lee, and W. Rhee.An image grid can be worth a video: Zero-shot video question answering using a vlm.arXiv preprint arXiv:2403.18406, 2024.
  • [25]P.-Y. Lab and T. A. etc.Open-sora-plan, Apr. 2024.
  • [26]B. Li, K. Zhang, H. Zhang, D. Guo, R. Zhang, F. Li, Y. Zhang, Z. Liu, and C. Li.Llava-next: Stronger llms supercharge multimodal capabilities in the wild, May 2024.
  • [27]B. Li, Y. Zhang, L. Chen, J. Wang, J. Yang, and Z. Liu.Otter: A multi-modal model with in-context instruction tuning.arXiv preprint arXiv:2305.03726, 2023.
  • [28]J. Li, D. Li, S. Savarese, and S. Hoi.Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models.arXiv preprint arXiv:2301.12597, 2023.
  • [29]K. Li, Y. He, Y. Wang, Y. Li, W. Wang, P. Luo, Y. Wang, L. Wang, and Y. Qiao.Videochat: Chat-centric video understanding.arXiv preprint arXiv:2305.06355, 2023.
  • [30]K. Li, Y. Wang, Y. He, Y. Li, Y. Wang, Y. Liu, Z. Wang, J. Xu, G. Chen, P. Luo, et al.Mvbench: A comprehensive multi-modal video understanding benchmark.arXiv preprint arXiv:2311.17005, 2023.
  • [31]Y. Li, Y. Song, L. Cao, J. Tetreault, L. Goldberg, A. Jaimes, and J. Luo.Tgif: A new dataset and benchmark on animated gif description.In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4641–4650, 2016.
  • [32]Y. Li, C. Wang, and J. Jia.Llama-vid: An image is worth 2 tokens in large language models.arXiv preprint arXiv:2311.17043, 2023.
  • [33]Z. Li, B. Yang, Q. Liu, Z. Ma, S. Zhang, J. Yang, Y. Sun, Y. Liu, and X. Bai.Monkey: Image resolution and text label are important things for large multi-modal models.arXiv preprint arXiv:2311.06607, 2023.
  • [34]B. Lin, B. Zhu, Y. Ye, M. Ning, P. Jin, and L. Yuan.Video-llava: Learning united visual representation by alignment before projection.arXiv preprint arXiv:2311.10122, 2023.
  • [35]H. Liu, C. Li, Y. Li, and Y. J. Lee.Improved baselines with visual instruction tuning.arXiv preprint arXiv:2310.03744, 2023.
  • [36]H. Liu, C. Li, Y. Li, B. Li, Y. Zhang, S. Shen, and Y. J. Lee.Llava-next: Improved reasoning, ocr, and world knowledge, January 2024.
  • [37]H. Liu, C. Li, Q. Wu, and Y. J. Lee.Visual instruction tuning.arXiv preprint arXiv:2304.08485, 2023.
  • [38]Y. Liu, S. Li, Y. Liu, Y. Wang, S. Ren, L. Li, S. Chen, X. Sun, and L. Hou.Tempcompass: Do video llms really understand videos?arXiv preprint arXiv:2403.00476, 2024.
  • [39]Y. Liu, K. Zhang, Y. Li, Z. Yan, C. Gao, R. Chen, Z. Yuan, Y. Huang, H. Sun, J. Gao, et al.Sora: A review on background, technology, limitations, and opportunities of large vision models.arXiv preprint arXiv:2402.17177, 2024.
  • [40]R. Luo, Z. Zhao, M. Yang, J. Dong, M. Qiu, P. Lu, T. Wang, and Z. Wei.Valley: Video assistant with large language model enhanced ability.arXiv preprint arXiv:2306.07207, 2023.
  • [41]X. Ma, Y. Wang, G. Jia, X. Chen, Z. Liu, Y.-F. Li, C. Chen, and Y. Qiao.Latte: Latent diffusion transformer for video generation.arXiv preprint arXiv:2401.03048, 2024.
  • [42]M. Maaz, H. Rasheed, S. Khan, and F. S. Khan.Video-chatgpt: Towards detailed video understanding via large vision and language models.arXiv preprint arXiv:2306.05424, 2023.
  • [43]mixkit.mixkit.https://mixkit.com/videos/, 2023.
  • [44]M. Ning, B. Zhu, Y. Xie, B. Lin, J. Cui, L. Yuan, D. Chen, and L. Yuan.Video-bench: A comprehensive benchmark and toolkit for evaluating video-based large language models.arXiv preprint arXiv:2311.16103, 2023.
  • [45]OpenAI.Chatgpt.https://chat.openai.com/, 2023.
  • [46]OpenAI.Gpt-4v(ision) system card.https://cdn.openai.com/papers/GPTV_System_Card.pdf, 2023.
  • [47]W. Peebles and S. Xie.Scalable diffusion models with transformers.In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4195–4205, 2023.
  • [48]Pexels.Pexels.https://www.pexels.com/videos/, 2023.
  • [49]pixabay.pixabay.https://pixabay.com/videos/, 2023.
  • [50]A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al.Learning transferable visual models from natural language supervision.In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  • [51]C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu.Exploring the limits of transfer learning with a unified text-to-text transformer.The Journal of Machine Learning Research, 21(1):5485–5551, 2020.
  • [52]R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer.High-resolution image synthesis with latent diffusion models.In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10684–10695, 2022.
  • [53]N. Ruiz, Y. Li, V. Jampani, Y. Pritch, M. Rubinstein, and K. Aberman.Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 22500–22510, 2023.
  • [54]U. Singer, A. Polyak, T. Hayes, X. Yin, J. An, S. Zhang, Q. Hu, H. Yang, O. Ashual, O. Gafni, et al.Make-a-video: Text-to-video generation without text-video data.arXiv preprint arXiv:2209.14792, 2022.
  • [55]J. Wang, L. Meng, Z. Weng, B. He, Z. Wu, and Y.-G. Jiang.To see is to believe: Prompting gpt-4v for better visual instruction tuning.arXiv preprint arXiv:2311.07574, 2023.
  • [56]W. Wang, Q. Lv, W. Yu, W. Hong, J. Qi, Y. Wang, J. Ji, Z. Yang, L. Zhao, X. Song, et al.Cogvlm: Visual expert for pretrained language models.arXiv preprint arXiv:2311.03079, 2023.
  • [57]Y. Wang, X. Chen, X. Ma, S. Zhou, Z. Huang, Y. Wang, C. Yang, Y. He, J. Yu, P. Yang, et al.Lavie: High-quality video generation with cascaded latent diffusion models.arXiv preprint arXiv:2309.15103, 2023.
  • [58]H. Wu, Z. Zhang, E. Zhang, C. Chen, L. Liao, A. Wang, C. Li, W. Sun, Q. Yan, G. Zhai, and W. Lin.Q-Bench: A benchmark for general-purpose foundation models on low-level vision.In ICLR, 2024.
  • [59]H. Wu, Z. Zhang, E. Zhang, C. Chen, L. Liao, A. Wang, K. Xu, C. Li, J. Hou, G. Zhai, et al.Q-Instruct: Improving low-level visual abilities for multi-modality foundation models.IEEE CVPR, pages 1–16, 2024.
  • [60]J. Z. Wu, Y. Ge, X. Wang, S. W. Lei, Y. Gu, Y. Shi, W. Hsu, Y. Shan, X. Qie, and M. Z. Shou.Tune-a-video: One-shot tuning of image diffusion models for text-to-video generation.In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7623–7633, 2023.
  • [61]J. Xiao, X. Shang, A. Yao, and T.-S. Chua.Next-qa: Next phase of question-answering to explaining temporal actions.In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9777–9786, 2021.
  • [62]D. Xu, Z. Zhao, J. Xiao, F. Wu, H. Zhang, X. He, and Y. Zhuang.Video question answering via gradually refined attention over appearance and motion.In Proceedings of the 25th ACM international conference on Multimedia, pages 1645–1653, 2017.
  • [63]Q. Ye, H. Xu, G. Xu, J. Ye, M. Yan, Y. Zhou, J. Wang, A. Hu, P. Shi, Y. Shi, et al.mplug-owl: Modularization empowers large language models with multimodality.arXiv preprint arXiv:2304.14178, 2023.
  • [64]K. Yi, C. Gan, Y. Li, P. Kohli, J. Wu, A. Torralba, and J. B. Tenenbaum.Clevrer: Collision events for video representation and reasoning.arXiv preprint arXiv:1910.01442, 2019.
  • [65]A. Young, B. Chen, C. Li, C. Huang, G. Zhang, G. Zhang, H. Li, J. Zhu, J. Chen, J. Chang, et al.Yi: Open foundation models by 01. ai.arXiv preprint arXiv:2403.04652, 2024.
  • [66]F. Yu, H. Chen, X. Wang, W. Xian, Y. Chen, F. Liu, V. Madhavan, and T. Darrell.Bdd100k: A diverse driving dataset for heterogeneous multitask learning.In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2636–2645, 2020.
  • [67]Z. Yu, D. Xu, J. Yu, T. Yu, Z. Zhao, Y. Zhuang, and D. Tao.Activitynet-qa: A dataset for understanding complex web videos via question answering.In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 9127–9134, 2019.
  • [68]Y. Zeng, G. Wei, J. Zheng, J. Zou, Y. Wei, Y. Zhang, and H. Li.Make pixels dance: High-dynamic video generation.arXiv preprint arXiv:2311.10982, 2023.
  • [69]H. Zhang, X. Li, and L. Bing.Video-llama: An instruction-tuned audio-visual language model for video understanding.arXiv preprint arXiv:2306.02858, 2023.
  • [70]L. Zhang, A. Rao, and M. Agrawala.Adding conditional control to text-to-image diffusion models.In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3836–3847, 2023.
  • [71]R. Zhang, J. Han, A. Zhou, X. Hu, S. Yan, P. Lu, H. Li, P. Gao, and Y. Qiao.Llama-adapter: Efficient fine-tuning of language models with zero-init attention.arXiv preprint arXiv:2303.16199, 2023.
  • [72]B. Zhou, Y. Hu, X. Weng, J. Jia, J. Luo, X. Liu, J. Wu, and L. Huang.Tinyllava: A framework of small-scale large multimodal models.arXiv preprint arXiv:2402.14289, 2024.
  • [73]D. Zhou, W. Wang, H. Yan, W. Lv, Y. Zhu, and J. Feng.Magicvideo: Efficient video generation with latent diffusion models.arXiv preprint arXiv:2211.11018, 2022.
  • [74]Y. Zhuang, Y. Rui, T. S. Huang, and S. Mehrotra.Adaptive key frame extraction using unsupervised clustering.In Proceedings 1998 international conference on image processing. icip98 (cat. no. 98cb36269), volume 1, pages 866–870. IEEE, 1998.
ShareGPT4Video: Improving Video Understanding and Generation with Better Captions (2024)

References

Top Articles
Latest Posts
Article information

Author: Melvina Ondricka

Last Updated:

Views: 5525

Rating: 4.8 / 5 (68 voted)

Reviews: 83% of readers found this page helpful

Author information

Name: Melvina Ondricka

Birthday: 2000-12-23

Address: Suite 382 139 Shaniqua Locks, Paulaborough, UT 90498

Phone: +636383657021

Job: Dynamic Government Specialist

Hobby: Kite flying, Watching movies, Knitting, Model building, Reading, Wood carving, Paintball

Introduction: My name is Melvina Ondricka, I am a helpful, fancy, friendly, innocent, outstanding, courageous, thoughtful person who loves writing and wants to share my knowledge and understanding with you.