Upload videollama3_original_model
Browse filesThis view is limited to 50 files because it contains too many changes. Β See raw diff
- videollama3_original_model/.gitattributes +35 -0
- videollama3_original_model/.huggingface/.gitignore +1 -0
- videollama3_original_model/.huggingface/download/.gitattributes.lock +0 -0
- videollama3_original_model/.huggingface/download/.gitattributes.metadata +3 -0
- videollama3_original_model/.huggingface/download/README.md.lock +0 -0
- videollama3_original_model/.huggingface/download/README.md.metadata +3 -0
- videollama3_original_model/.huggingface/download/added_tokens.json.lock +0 -0
- videollama3_original_model/.huggingface/download/added_tokens.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/chat_template.json.lock +0 -0
- videollama3_original_model/.huggingface/download/chat_template.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/config.json.lock +0 -0
- videollama3_original_model/.huggingface/download/config.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/configuration_videollama3.py.lock +0 -0
- videollama3_original_model/.huggingface/download/configuration_videollama3.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.lock +0 -0
- videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/generation_config.json.lock +0 -0
- videollama3_original_model/.huggingface/download/generation_config.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/image_processing_videollama3.py.lock +0 -0
- videollama3_original_model/.huggingface/download/image_processing_videollama3.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/merges.txt.lock +0 -0
- videollama3_original_model/.huggingface/download/merges.txt.metadata +3 -0
- videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.lock +0 -0
- videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.metadata +3 -0
- videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.lock +0 -0
- videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.metadata +3 -0
- videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.lock +0 -0
- videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.metadata +3 -0
- videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.lock +0 -0
- videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.metadata +3 -0
- videollama3_original_model/.huggingface/download/model.safetensors.index.json.lock +0 -0
- videollama3_original_model/.huggingface/download/model.safetensors.index.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/modeling_videollama3.py.lock +0 -0
- videollama3_original_model/.huggingface/download/modeling_videollama3.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.lock +0 -0
- videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/preprocessor_config.json.lock +0 -0
- videollama3_original_model/.huggingface/download/preprocessor_config.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/processing_videollama3.py.lock +0 -0
- videollama3_original_model/.huggingface/download/processing_videollama3.py.metadata +3 -0
- videollama3_original_model/.huggingface/download/processor_config.json.lock +0 -0
- videollama3_original_model/.huggingface/download/processor_config.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/special_tokens_map.json.lock +0 -0
- videollama3_original_model/.huggingface/download/special_tokens_map.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/tokenizer_config.json.lock +0 -0
- videollama3_original_model/.huggingface/download/tokenizer_config.json.metadata +3 -0
- videollama3_original_model/.huggingface/download/vocab.json.lock +0 -0
- videollama3_original_model/.huggingface/download/vocab.json.metadata +3 -0
- videollama3_original_model/README.md +138 -0
- videollama3_original_model/added_tokens.json +27 -0
videollama3_original_model/.gitattributes
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
videollama3_original_model/.huggingface/.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
*
|
videollama3_original_model/.huggingface/download/.gitattributes.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/.gitattributes.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
a6344aac8c09253b3b630fb776ae94478aa0275b
|
| 3 |
+
1763248024.9122677
|
videollama3_original_model/.huggingface/download/README.md.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/README.md.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
c3a6ec2b1e4a9b96d16fb9693d66448828c9ca28
|
| 3 |
+
1763248024.9312024
|
videollama3_original_model/.huggingface/download/added_tokens.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/added_tokens.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
066b1421f76beca53bde9481638b4826054c122f
|
| 3 |
+
1763248024.8180633
|
videollama3_original_model/.huggingface/download/chat_template.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/chat_template.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
f67d31e44cc5c21bd8f1b7b26d27417e98691e6b
|
| 3 |
+
1763248024.8308272
|
videollama3_original_model/.huggingface/download/config.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/config.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
a5b54001faaa818772e4ab6274e95383186c431a
|
| 3 |
+
1763248024.7803109
|
videollama3_original_model/.huggingface/download/configuration_videollama3.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/configuration_videollama3.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
70cc7c7620eb451ae84d4bb79a6e0845a6c9b4df
|
| 3 |
+
1763248024.749131
|
videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/configuration_videollama3_encoder.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
9bd0decacc5c352b38c3249d1556dd91c6848fa8
|
| 3 |
+
1763248024.796934
|
videollama3_original_model/.huggingface/download/generation_config.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/generation_config.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
2e995d693e130e3e92b87ff760bbe99ccdc54ea0
|
| 3 |
+
1763248024.7491724
|
videollama3_original_model/.huggingface/download/image_processing_videollama3.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/image_processing_videollama3.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
91099b890869d0a0b1834a313fc6d107bdd8ab3f
|
| 3 |
+
1763248024.8219907
|
videollama3_original_model/.huggingface/download/merges.txt.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/merges.txt.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
31349551d90c7606f325fe0f11bbb8bd5fa0d7c7
|
| 3 |
+
1763248024.8034952
|
videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/model-00001-of-00004.safetensors.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
cf34749d69dab7940984ef88be8a33e161e467809c0aacf8825ac7de0a841cf6
|
| 3 |
+
1763248029.6650863
|
videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/model-00002-of-00004.safetensors.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
0ff2ad8aea2d46c9a2f215a781975144a09de6c1d9a8225d71da130913c34f4a
|
| 3 |
+
1763248029.674504
|
videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/model-00003-of-00004.safetensors.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
bd3e1c5410fb5b59e26fab11ea36d594a8ff0e11ff9fa56853c8f50a9ec3f61f
|
| 3 |
+
1763248029.780972
|
videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/model-00004-of-00004.safetensors.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
47e9f7d43163f99f20820ae8f6a95c8440cc66d195b75aa6a662372cacfc9bfb
|
| 3 |
+
1763248026.0995622
|
videollama3_original_model/.huggingface/download/model.safetensors.index.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/model.safetensors.index.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
596eaa3c7d781aeab36fc27a7f396827be8b4ca6
|
| 3 |
+
1763248024.8861187
|
videollama3_original_model/.huggingface/download/modeling_videollama3.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/modeling_videollama3.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
1d72a96961bef654b467a0bbfb9bab20e3bad338
|
| 3 |
+
1763248024.8835113
|
videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/modeling_videollama3_encoder.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
8ce6f86eb9763de6ad86f66f94aea99a6d35412d
|
| 3 |
+
1763248024.950885
|
videollama3_original_model/.huggingface/download/preprocessor_config.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/preprocessor_config.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
067f85616904a5313af095dfde820ee42654763a
|
| 3 |
+
1763248024.9201517
|
videollama3_original_model/.huggingface/download/processing_videollama3.py.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/processing_videollama3.py.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
0f3b28b868e9616afa3ef47fc58994ce4c89d135
|
| 3 |
+
1763248024.9481933
|
videollama3_original_model/.huggingface/download/processor_config.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/processor_config.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
379794013adc490b089027c06a72663e4c0c6d2d
|
| 3 |
+
1763248024.9543974
|
videollama3_original_model/.huggingface/download/special_tokens_map.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/special_tokens_map.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
ac23c0aaa2434523c494330aeb79c58395378103
|
| 3 |
+
1763248024.9825387
|
videollama3_original_model/.huggingface/download/tokenizer_config.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/tokenizer_config.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
c765d2ec9fcd7008d21010e9354275da5312d710
|
| 3 |
+
1763248024.977651
|
videollama3_original_model/.huggingface/download/vocab.json.lock
ADDED
|
File without changes
|
videollama3_original_model/.huggingface/download/vocab.json.metadata
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
d5b763e368861e7f5096e7ff1b49f92fbccf8ae6
|
| 2 |
+
6bce3a0a3866c4791a74d83d78f6824c3af64ec3
|
| 3 |
+
1763248025.0156815
|
videollama3_original_model/README.md
ADDED
|
@@ -0,0 +1,138 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
tags:
|
| 4 |
+
- multi-modal
|
| 5 |
+
- large-language-model
|
| 6 |
+
- video-language-model
|
| 7 |
+
license: apache-2.0
|
| 8 |
+
datasets:
|
| 9 |
+
- lmms-lab/LLaVA-OneVision-Data
|
| 10 |
+
- allenai/pixmo-docs
|
| 11 |
+
- HuggingFaceM4/Docmatix
|
| 12 |
+
- lmms-lab/LLaVA-Video-178K
|
| 13 |
+
- ShareGPT4Video/ShareGPT4Video
|
| 14 |
+
language:
|
| 15 |
+
- en
|
| 16 |
+
metrics:
|
| 17 |
+
- accuracy
|
| 18 |
+
pipeline_tag: video-text-to-text
|
| 19 |
+
base_model:
|
| 20 |
+
- Qwen/Qwen2.5-7B-Instruct
|
| 21 |
+
- DAMO-NLP-SG/VideoLLaMA3-7B-Image
|
| 22 |
+
---
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
<p align="center">
|
| 26 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/626938b16f8f86ad21deb989/tt5KYnAUmQlHtfB1-Zisl.png" width="150" style="margin-bottom: 0.2;"/>
|
| 27 |
+
<p>
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
<h3 align="center"><a href="https://arxiv.org/abs/2501.13106">VideoLLaMA 3: Frontier Multimodal Foundation Models for Video Understanding</a></h3>
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
<h5 align="center"> If you like our project, please give us a star β on <a href="https://github.com/DAMO-NLP-SG/VideoLLaMA3">Github</a> for the latest update. </h5>
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
## π° News
|
| 37 |
+
<!-- * **[2024.01.23]** ππ Update technical report. If you have works closely related to VideoLLaMA3 but not mentioned in the paper, feel free to let us know. -->
|
| 38 |
+
* **[2024.01.24]** π₯π₯ Online Demo is available: [VideoLLaMA3-Image-7B](https://huggingface.co/spaces/lixin4ever/VideoLLaMA3-Image), [VideoLLaMA3-7B](https://huggingface.co/spaces/lixin4ever/VideoLLaMA3).
|
| 39 |
+
* **[2024.01.22]** Release models and inference code of VideoLLaMA 3.
|
| 40 |
+
|
| 41 |
+
## π Introduction
|
| 42 |
+
VideoLLaMA 3 represents a state-of-the-art series of multimodal foundation models designed to excel in both image and video understanding tasks. Leveraging advanced architectures, VideoLLaMA 3 demonstrates exceptional capabilities in processing and interpreting visual content across various contexts. These models are specifically designed to address complex multimodal challenges, such as integrating textual and visual information, extracting insights from sequential video data, and performing high-level reasoning over both dynamic and static visual scenes.
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
## π Model Zoo
|
| 49 |
+
| Model | Base Model | HF Link |
|
| 50 |
+
| -------------------- | ------------ | ------------------------------------------------------------ |
|
| 51 |
+
| VideoLLaMA3-7B (**This Checkpoint**) | Qwen2.5-7B | [DAMO-NLP-SG/VideoLLaMA3-7B](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-7B) |
|
| 52 |
+
| VideoLLaMA3-2B | Qwen2.5-1.5B | [DAMO-NLP-SG/VideoLLaMA3-2B](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-2B) |
|
| 53 |
+
| VideoLLaMA3-7B-Image | Qwen2.5-7B | [DAMO-NLP-SG/VideoLLaMA3-7B-Image](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-7B-Image) |
|
| 54 |
+
| VideoLLaMA3-2B-Image | Qwen2.5-1.5B | [DAMO-NLP-SG/VideoLLaMA3-2B-Image](https://huggingface.co/DAMO-NLP-SG/VideoLLaMA3-2B-Image) |
|
| 55 |
+
|
| 56 |
+
We also upload the tuned vision encoder of VideoLLaMA3-7B for wider application:
|
| 57 |
+
|
| 58 |
+
| Model | Base Model | HF Link |
|
| 59 |
+
| ----------------------------- | ------------------------- | ------------------------------------------------------------ |
|
| 60 |
+
| VideoLLaMA3-7B Vision Encoder | siglip-so400m-patch14-384 | [DAMO-NLP-SG/VL3-SigLIP-NaViT](https://huggingface.co/DAMO-NLP-SG/VL3-SigLIP-NaViT) |
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
## π Main Results
|
| 65 |
+
|
| 66 |
+
|
| 67 |
+
<img width="500" alt="image" src="https://cdn-uploads.huggingface.co/production/uploads/609115c79a8bcaa437b234a9/ggmVF_v68QsHuPFVZR3MR.png">
|
| 68 |
+
|
| 69 |
+
* \* denotes the reproduced results.
|
| 70 |
+
|
| 71 |
+
## π€ Quick Start
|
| 72 |
+
```python
|
| 73 |
+
import torch
|
| 74 |
+
from transformers import AutoModelForCausalLM, AutoProcessor, AutoModel, AutoImageProcessor
|
| 75 |
+
|
| 76 |
+
model_name = "DAMO-NLP-SG/VideoLLaMA3-7B"
|
| 77 |
+
|
| 78 |
+
model = AutoModelForCausalLM.from_pretrained(
|
| 79 |
+
model_name,
|
| 80 |
+
trust_remote_code=True,
|
| 81 |
+
device_map="auto",
|
| 82 |
+
torch_dtype=torch.bfloat16,
|
| 83 |
+
attn_implementation="flash_attention_2",
|
| 84 |
+
)
|
| 85 |
+
processor = AutoProcessor.from_pretrained(model_name, trust_remote_code=True)
|
| 86 |
+
video_path = "put your video path here"
|
| 87 |
+
question = "Describe this video in detail."
|
| 88 |
+
|
| 89 |
+
# Video conversation
|
| 90 |
+
conversation = [
|
| 91 |
+
{"role": "system", "content": "You are a helpful assistant."},
|
| 92 |
+
{
|
| 93 |
+
"role": "user",
|
| 94 |
+
"content": [
|
| 95 |
+
{"type": "video", "video": {"video_path": video_path, "fps": 1, "max_frames": 128}},
|
| 96 |
+
{"type": "text", "text": question},
|
| 97 |
+
]
|
| 98 |
+
},
|
| 99 |
+
]
|
| 100 |
+
|
| 101 |
+
inputs = processor(conversation=conversation, return_tensors="pt")
|
| 102 |
+
inputs = {k: v.cuda() if isinstance(v, torch.Tensor) else v for k, v in inputs.items()}
|
| 103 |
+
if "pixel_values" in inputs:
|
| 104 |
+
inputs["pixel_values"] = inputs["pixel_values"].to(torch.bfloat16)
|
| 105 |
+
output_ids = model.generate(**inputs, max_new_tokens=128)
|
| 106 |
+
response = processor.batch_decode(output_ids, skip_special_tokens=True)[0].strip()
|
| 107 |
+
print(response)
|
| 108 |
+
```
|
| 109 |
+
|
| 110 |
+
|
| 111 |
+
## Citation
|
| 112 |
+
|
| 113 |
+
If you find VideoLLaMA useful for your research and applications, please cite using this BibTeX:
|
| 114 |
+
```bibtex
|
| 115 |
+
@article{damonlpsg2025videollama3,
|
| 116 |
+
title={VideoLLaMA 3: Frontier Multimodal Foundation Models for Image and Video Understanding},
|
| 117 |
+
author={Boqiang Zhang, Kehan Li, Zesen Cheng, Zhiqiang Hu, Yuqian Yuan, Guanzheng Chen, Sicong Leng, Yuming Jiang, Hang Zhang, Xin Li, Peng Jin, Wenqi Zhang, Fan Wang, Lidong Bing, Deli Zhao},
|
| 118 |
+
journal={arXiv preprint arXiv:2501.13106},
|
| 119 |
+
year={2025},
|
| 120 |
+
url = {https://arxiv.org/abs/2501.13106}
|
| 121 |
+
}
|
| 122 |
+
|
| 123 |
+
@article{damonlpsg2024videollama2,
|
| 124 |
+
title={VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs},
|
| 125 |
+
author={Cheng, Zesen and Leng, Sicong and Zhang, Hang and Xin, Yifei and Li, Xin and Chen, Guanzheng and Zhu, Yongxin and Zhang, Wenqi and Luo, Ziyang and Zhao, Deli and Bing, Lidong},
|
| 126 |
+
journal={arXiv preprint arXiv:2406.07476},
|
| 127 |
+
year={2024},
|
| 128 |
+
url = {https://arxiv.org/abs/2406.07476}
|
| 129 |
+
}
|
| 130 |
+
|
| 131 |
+
@article{damonlpsg2023videollama,
|
| 132 |
+
title = {Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding},
|
| 133 |
+
author = {Zhang, Hang and Li, Xin and Bing, Lidong},
|
| 134 |
+
journal = {arXiv preprint arXiv:2306.02858},
|
| 135 |
+
year = {2023},
|
| 136 |
+
url = {https://arxiv.org/abs/2306.02858}
|
| 137 |
+
}
|
| 138 |
+
```
|
videollama3_original_model/added_tokens.json
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"</tool_call>": 151658,
|
| 3 |
+
"<image>": 151665,
|
| 4 |
+
"<tool_call>": 151657,
|
| 5 |
+
"<|box_end|>": 151649,
|
| 6 |
+
"<|box_start|>": 151648,
|
| 7 |
+
"<|endoftext|>": 151643,
|
| 8 |
+
"<|file_sep|>": 151664,
|
| 9 |
+
"<|fim_middle|>": 151660,
|
| 10 |
+
"<|fim_pad|>": 151662,
|
| 11 |
+
"<|fim_prefix|>": 151659,
|
| 12 |
+
"<|fim_suffix|>": 151661,
|
| 13 |
+
"<|im_end|>": 151645,
|
| 14 |
+
"<|im_start|>": 151644,
|
| 15 |
+
"<|image_pad|>": 151655,
|
| 16 |
+
"<|object_ref_end|>": 151647,
|
| 17 |
+
"<|object_ref_start|>": 151646,
|
| 18 |
+
"<|quad_end|>": 151651,
|
| 19 |
+
"<|quad_start|>": 151650,
|
| 20 |
+
"<|repo_name|>": 151663,
|
| 21 |
+
"<|stream_end|>": 151667,
|
| 22 |
+
"<|stream_start|>": 151666,
|
| 23 |
+
"<|video_pad|>": 151656,
|
| 24 |
+
"<|vision_end|>": 151653,
|
| 25 |
+
"<|vision_pad|>": 151654,
|
| 26 |
+
"<|vision_start|>": 151652
|
| 27 |
+
}
|