KDliang commited on
Commit
9023b2e
Β·
verified Β·
1 Parent(s): 9c18911

Upload videollama3_original_model

Browse files
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. videollama3_original_model/.gitattributes +35 -0
  2. videollama3_original_model/.huggingface/.gitignore +1 -0
  3. videollama3_original_model/.huggingface/download/.gitattributes.lock +0 -0
  4. videollama3_original_model/.huggingface/download/.gitattributes.metadata +3 -0
  5. videollama3_original_model/.huggingface/download/README.md.lock +0 -0
  6. videollama3_original_model/.huggingface/download/README.md.metadata +3 -0
  7. videollama3_original_model/.huggingface/download/added_tokens.json.lock +0 -0
  8. videollama3_original_model/.huggingface/download/added_tokens.json.metadata +3 -0
  9. videollama3_original_model/.huggingface/download/chat_template.json.lock +0 -0
  10. videollama3_original_model/.huggingface/download/chat_template.json.metadata +3 -0
  11. videollama3_original_model/.huggingface/download/config.json.lock +0 -0
  12. videollama3_original_model/.huggingface/download/config.json.metadata +3 -0
  13. videollama3_original_model/.huggingface/download/configuration_videollama3.py.lock +0 -0
  14. videollama3_original_model/.huggingface/download/configuration_videollama3.py.metadata +3 -0
  15. videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.lock +0 -0
  16. videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.metadata +3 -0
  17. videollama3_original_model/.huggingface/download/generation_config.json.lock +0 -0
  18. videollama3_original_model/.huggingface/download/generation_config.json.metadata +3 -0
  19. videollama3_original_model/.huggingface/download/image_processing_videollama3.py.lock +0 -0
  20. videollama3_original_model/.huggingface/download/image_processing_videollama3.py.metadata +3 -0
  21. videollama3_original_model/.huggingface/download/merges.txt.lock +0 -0
  22. videollama3_original_model/.huggingface/download/merges.txt.metadata +3 -0
  23. videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.lock +0 -0
  24. videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.metadata +3 -0
  25. videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.lock +0 -0
  26. videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.metadata +3 -0
  27. videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.lock +0 -0
  28. videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.metadata +3 -0
  29. videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.lock +0 -0
  30. videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.metadata +3 -0
  31. videollama3_original_model/.huggingface/download/model.safetensors.index.json.lock +0 -0
  32. videollama3_original_model/.huggingface/download/model.safetensors.index.json.metadata +3 -0
  33. videollama3_original_model/.huggingface/download/modeling_videollama3.py.lock +0 -0
  34. videollama3_original_model/.huggingface/download/modeling_videollama3.py.metadata +3 -0
  35. videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.lock +0 -0
  36. videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.metadata +3 -0
  37. videollama3_original_model/.huggingface/download/preprocessor_config.json.lock +0 -0
  38. videollama3_original_model/.huggingface/download/preprocessor_config.json.metadata +3 -0
  39. videollama3_original_model/.huggingface/download/processing_videollama3.py.lock +0 -0
  40. videollama3_original_model/.huggingface/download/processing_videollama3.py.metadata +3 -0
  41. videollama3_original_model/.huggingface/download/processor_config.json.lock +0 -0
  42. videollama3_original_model/.huggingface/download/processor_config.json.metadata +3 -0
  43. videollama3_original_model/.huggingface/download/special_tokens_map.json.lock +0 -0
  44. videollama3_original_model/.huggingface/download/special_tokens_map.json.metadata +3 -0
  45. videollama3_original_model/.huggingface/download/tokenizer_config.json.lock +0 -0
  46. videollama3_original_model/.huggingface/download/tokenizer_config.json.metadata +3 -0
  47. videollama3_original_model/.huggingface/download/vocab.json.lock +0 -0
  48. videollama3_original_model/.huggingface/download/vocab.json.metadata +3 -0
  49. videollama3_original_model/README.md +138 -0
  50. videollama3_original_model/added_tokens.json +27 -0
videollama3_original_model/.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
videollama3_original_model/.huggingface/.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ *
videollama3_original_model/.huggingface/download/.gitattributes.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/.gitattributes.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ a6344aac8c09253b3b630fb776ae94478aa0275b
3
+ 1763248024.9122677
videollama3_original_model/.huggingface/download/README.md.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/README.md.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ c3a6ec2b1e4a9b96d16fb9693d66448828c9ca28
3
+ 1763248024.9312024
videollama3_original_model/.huggingface/download/added_tokens.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/added_tokens.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 066b1421f76beca53bde9481638b4826054c122f
3
+ 1763248024.8180633
videollama3_original_model/.huggingface/download/chat_template.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/chat_template.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ f67d31e44cc5c21bd8f1b7b26d27417e98691e6b
3
+ 1763248024.8308272
videollama3_original_model/.huggingface/download/config.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/config.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ a5b54001faaa818772e4ab6274e95383186c431a
3
+ 1763248024.7803109
videollama3_original_model/.huggingface/download/configuration_videollama3.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/configuration_videollama3.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 70cc7c7620eb451ae84d4bb79a6e0845a6c9b4df
3
+ 1763248024.749131
videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 9bd0decacc5c352b38c3249d1556dd91c6848fa8
3
+ 1763248024.796934
videollama3_original_model/.huggingface/download/generation_config.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/generation_config.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 2e995d693e130e3e92b87ff760bbe99ccdc54ea0
3
+ 1763248024.7491724
videollama3_original_model/.huggingface/download/image_processing_videollama3.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/image_processing_videollama3.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 91099b890869d0a0b1834a313fc6d107bdd8ab3f
3
+ 1763248024.8219907
videollama3_original_model/.huggingface/download/merges.txt.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/merges.txt.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 31349551d90c7606f325fe0f11bbb8bd5fa0d7c7
3
+ 1763248024.8034952
videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ cf34749d69dab7940984ef88be8a33e161e467809c0aacf8825ac7de0a841cf6
3
+ 1763248029.6650863
videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 0ff2ad8aea2d46c9a2f215a781975144a09de6c1d9a8225d71da130913c34f4a
3
+ 1763248029.674504
videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ bd3e1c5410fb5b59e26fab11ea36d594a8ff0e11ff9fa56853c8f50a9ec3f61f
3
+ 1763248029.780972
videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 47e9f7d43163f99f20820ae8f6a95c8440cc66d195b75aa6a662372cacfc9bfb
3
+ 1763248026.0995622
videollama3_original_model/.huggingface/download/model.safetensors.index.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/model.safetensors.index.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 596eaa3c7d781aeab36fc27a7f396827be8b4ca6
3
+ 1763248024.8861187
videollama3_original_model/.huggingface/download/modeling_videollama3.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/modeling_videollama3.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 1d72a96961bef654b467a0bbfb9bab20e3bad338
3
+ 1763248024.8835113
videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 8ce6f86eb9763de6ad86f66f94aea99a6d35412d
3
+ 1763248024.950885
videollama3_original_model/.huggingface/download/preprocessor_config.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/preprocessor_config.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 067f85616904a5313af095dfde820ee42654763a
3
+ 1763248024.9201517
videollama3_original_model/.huggingface/download/processing_videollama3.py.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/processing_videollama3.py.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 0f3b28b868e9616afa3ef47fc58994ce4c89d135
3
+ 1763248024.9481933
videollama3_original_model/.huggingface/download/processor_config.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/processor_config.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 379794013adc490b089027c06a72663e4c0c6d2d
3
+ 1763248024.9543974
videollama3_original_model/.huggingface/download/special_tokens_map.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/special_tokens_map.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ ac23c0aaa2434523c494330aeb79c58395378103
3
+ 1763248024.9825387
videollama3_original_model/.huggingface/download/tokenizer_config.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/tokenizer_config.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ c765d2ec9fcd7008d21010e9354275da5312d710
3
+ 1763248024.977651
videollama3_original_model/.huggingface/download/vocab.json.lock ADDED
File without changes
videollama3_original_model/.huggingface/download/vocab.json.metadata ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
2
+ 6bce3a0a3866c4791a74d83d78f6824c3af64ec3
3
+ 1763248025.0156815
videollama3_original_model/README.md ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - multi-modal
5
+ - large-language-model
6
+ - video-language-model
7
+ license: apache-2.0
8
+ datasets:
9
+ - lmms-lab/LLaVA-OneVision-Data
10
+ - allenai/pixmo-docs
11
+ - HuggingFaceM4/Docmatix
12
+ - lmms-lab/LLaVA-Video-178K
13
+ - ShareGPT4Video/ShareGPT4Video
14
+ language:
15
+ - en
16
+ metrics:
17
+ - accuracy
18
+ pipeline_tag: video-text-to-text
19
+ base_model:
20
+ - Qwen/Qwen2.5-7B-Instruct
21
+ - DAMO-NLP-SG/VideoLLaMA3-7B-Image
22
+ ---
23
+
24
+
25
+ <p align="center">
26
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/626938b16f8f86ad21deb989/tt5KYnAUmQlHtfB1-Zisl.png" width="150" style="margin-bottom: 0.2;"/>
27
+ <p>
28
+
29
+
30
+ <h3 align="center"><a href="https://arxiv.org/abs/2501.13106">VideoLLaMA 3: Frontier Multimodal Foundation Models for Video Understanding</a></h3>
31
+
32
+
33
+ <h5 align="center"> If you like our project, please give us a star ⭐ on <a href="https://github.com/DAMO-NLP-SG/VideoLLaMA3">Github</a> for the latest update. </h5>
34
+
35
+
36
+ ## πŸ“° News
37
+ <!-- * **[2024.01.23]** πŸ‘‹πŸ‘‹ Update technical report. If you have works closely related to VideoLLaMA3 but not mentioned in the paper, feel free to let us know. -->
38
+ * **[2024.01.24]** πŸ”₯πŸ”₯ Online Demo is available: [VideoLLaMA3-Image-7B](https://huggingface.co/spaces/lixin4ever/VideoLLaMA3-Image), [VideoLLaMA3-7B](https://huggingface.co/spaces/lixin4ever/VideoLLaMA3).
39
+ * **[2024.01.22]** Release models and inference code of VideoLLaMA 3.
40
+
41
+ ## 🌟 Introduction
42
+ VideoLLaMA 3 represents a state-of-the-art series of multimodal foundation models designed to excel in both image and video understanding tasks. Leveraging advanced architectures, VideoLLaMA 3 demonstrates exceptional capabilities in processing and interpreting visual content across various contexts. These models are specifically designed to address complex multimodal challenges, such as integrating textual and visual information, extracting insights from sequential video data, and performing high-level reasoning over both dynamic and static visual scenes.
43
+
44
+
45
+
46
+
47
+
48
+ ## 🌎 Model Zoo
49
+ | Model | Base Model | HF Link |
50
+ | -------------------- | ------------ | ------------------------------------------------------------ |
51
+ | VideoLLaMA3-7B (**This Checkpoint**) | Qwen2.5-7B | [DAMO-NLP-SG/VideoLLaMA3-7B](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-7B) |
52
+ | VideoLLaMA3-2B | Qwen2.5-1.5B | [DAMO-NLP-SG/VideoLLaMA3-2B](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-2B) |
53
+ | VideoLLaMA3-7B-Image | Qwen2.5-7B | [DAMO-NLP-SG/VideoLLaMA3-7B-Image](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-7B-Image) |
54
+ | VideoLLaMA3-2B-Image | Qwen2.5-1.5B | [DAMO-NLP-SG/VideoLLaMA3-2B-Image](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-2B-Image) |
55
+
56
+ We also upload the tuned vision encoder of VideoLLaMA3-7B for wider application:
57
+
58
+ | Model | Base Model | HF Link |
59
+ | ----------------------------- | ------------------------- | ------------------------------------------------------------ |
60
+ | VideoLLaMA3-7B Vision Encoder | siglip-so400m-patch14-384 | [DAMO-NLP-SG/VL3-SigLIP-NaViT](https://huggingface.co/DAMO-NLP-SG/VL3-SigLIP-NaViT) |
61
+
62
+
63
+
64
+ ## πŸš€ Main Results
65
+
66
+
67
+ <img width="500" alt="image" src="https://cdn-uploads.huggingface.co/production/uploads/609115c79a8bcaa437b234a9/ggmVF_v68QsHuPFVZR3MR.png">
68
+
69
+ * \* denotes the reproduced results.
70
+
71
+ ## πŸ€– Quick Start
72
+ ```python
73
+ import torch
74
+ from transformers import AutoModelForCausalLM, AutoProcessor, AutoModel, AutoImageProcessor
75
+
76
+ model_name = "DAMO-NLP-SG/VideoLLaMA3-7B"
77
+
78
+ model = AutoModelForCausalLM.from_pretrained(
79
+ model_name,
80
+ trust_remote_code=True,
81
+ device_map="auto",
82
+ torch_dtype=torch.bfloat16,
83
+ attn_implementation="flash_attention_2",
84
+ )
85
+ processor = AutoProcessor.from_pretrained(model_name, trust_remote_code=True)
86
+ video_path = "put your video path here"
87
+ question = "Describe this video in detail."
88
+
89
+ # Video conversation
90
+ conversation = [
91
+ {"role": "system", "content": "You are a helpful assistant."},
92
+ {
93
+ "role": "user",
94
+ "content": [
95
+ {"type": "video", "video": {"video_path": video_path, "fps": 1, "max_frames": 128}},
96
+ {"type": "text", "text": question},
97
+ ]
98
+ },
99
+ ]
100
+
101
+ inputs = processor(conversation=conversation, return_tensors="pt")
102
+ inputs = {k: v.cuda() if isinstance(v, torch.Tensor) else v for k, v in inputs.items()}
103
+ if "pixel_values" in inputs:
104
+ inputs["pixel_values"] = inputs["pixel_values"].to(torch.bfloat16)
105
+ output_ids = model.generate(**inputs, max_new_tokens=128)
106
+ response = processor.batch_decode(output_ids, skip_special_tokens=True)[0].strip()
107
+ print(response)
108
+ ```
109
+
110
+
111
+ ## Citation
112
+
113
+ If you find VideoLLaMA useful for your research and applications, please cite using this BibTeX:
114
+ ```bibtex
115
+ @article{damonlpsg2025videollama3,
116
+ title={VideoLLaMA 3: Frontier Multimodal Foundation Models for Image and Video Understanding},
117
+ author={Boqiang Zhang, Kehan Li, Zesen Cheng, Zhiqiang Hu, Yuqian Yuan, Guanzheng Chen, Sicong Leng, Yuming Jiang, Hang Zhang, Xin Li, Peng Jin, Wenqi Zhang, Fan Wang, Lidong Bing, Deli Zhao},
118
+ journal={arXiv preprint arXiv:2501.13106},
119
+ year={2025},
120
+ url = {https://arxiv.org/abs/2501.13106}
121
+ }
122
+
123
+ @article{damonlpsg2024videollama2,
124
+ title={VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs},
125
+ author={Cheng, Zesen and Leng, Sicong and Zhang, Hang and Xin, Yifei and Li, Xin and Chen, Guanzheng and Zhu, Yongxin and Zhang, Wenqi and Luo, Ziyang and Zhao, Deli and Bing, Lidong},
126
+ journal={arXiv preprint arXiv:2406.07476},
127
+ year={2024},
128
+ url = {https://arxiv.org/abs/2406.07476}
129
+ }
130
+
131
+ @article{damonlpsg2023videollama,
132
+ title = {Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding},
133
+ author = {Zhang, Hang and Li, Xin and Bing, Lidong},
134
+ journal = {arXiv preprint arXiv:2306.02858},
135
+ year = {2023},
136
+ url = {https://arxiv.org/abs/2306.02858}
137
+ }
138
+ ```
videollama3_original_model/added_tokens.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<image>": 151665,
4
+ "<tool_call>": 151657,
5
+ "<|box_end|>": 151649,
6
+ "<|box_start|>": 151648,
7
+ "<|endoftext|>": 151643,
8
+ "<|file_sep|>": 151664,
9
+ "<|fim_middle|>": 151660,
10
+ "<|fim_pad|>": 151662,
11
+ "<|fim_prefix|>": 151659,
12
+ "<|fim_suffix|>": 151661,
13
+ "<|im_end|>": 151645,
14
+ "<|im_start|>": 151644,
15
+ "<|image_pad|>": 151655,
16
+ "<|object_ref_end|>": 151647,
17
+ "<|object_ref_start|>": 151646,
18
+ "<|quad_end|>": 151651,
19
+ "<|quad_start|>": 151650,
20
+ "<|repo_name|>": 151663,
21
+ "<|stream_end|>": 151667,
22
+ "<|stream_start|>": 151666,
23
+ "<|video_pad|>": 151656,
24
+ "<|vision_end|>": 151653,
25
+ "<|vision_pad|>": 151654,
26
+ "<|vision_start|>": 151652
27
+ }