Image-to-Video
Diffusers
wruisi commited on
Commit
2b16414
·
1 Parent(s): 3a1d880

Upload model

Browse files
Files changed (2) hide show
  1. README.md +223 -3
  2. lora.safetensors +3 -0
README.md CHANGED
@@ -1,3 +1,223 @@
1
- ---
2
- license: cc
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Lightricks/LTX-2.3
4
+ library_name: diffusers
5
+ license: apache-2.0
6
+ pipeline_tag: image-to-video
7
+ ---
8
+ # VBVR: A Very Big Video Reasoning Suite
9
+
10
+ <a href="https://video-reason.com" target="_blank">
11
+ <img alt="Project Page" src="https://img.shields.io/badge/Project%20-%20Homepage-4285F4" height="20" />
12
+ </a>
13
+ <a href="https://github.com/Video-Reason/VBVR-EvalKit" target="_blank">
14
+ <img alt="Code" src="https://img.shields.io/badge/Evaluation_code-VBVR_Bench-100000?style=flat-square&logo=github&logoColor=white" height="20" />
15
+ </a>
16
+ <a href="https://github.com/Video-Reason/VBVR-Wan2.2" target="_blank">
17
+ <img alt="Code" src="https://img.shields.io/badge/Training_code-VBVR_Wan2.2-100000?style=flat-square&logo=github&logoColor=white" height="20" />
18
+ </a>
19
+ <a href="https://github.com/Video-Reason/VBVR-DataFactory" target="_blank">
20
+ <img alt="Code" src="https://img.shields.io/badge/Data_code-VBVR_DataFactory-100000?style=flat-square&logo=github&logoColor=white" height="20" />
21
+ </a>
22
+ <a href="https://huggingface.co/papers/2602.20159" target="_blank">
23
+ <img alt="arXiv" src="https://img.shields.io/badge/arXiv-VBVR-red?logo=arxiv" height="20" />
24
+ </a>
25
+ <a href="https://huggingface.co/datasets/Video-Reason/VBVR-Dataset" target="_blank">
26
+ <img alt="Dataset" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Dataset-Data-ffc107?color=ffc107&logoColor=white" height="20" />
27
+ </a>
28
+ <a href="https://huggingface.co/datasets/Video-Reason/VBVR-Bench-Data" target="_blank">
29
+ <img alt="Bench Data" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Bench-Data-ffc107?color=ffc107&logoColor=white" height="20" />
30
+ </a>
31
+ <a href="https://huggingface.co/spaces/Video-Reason/VBVR-Bench-Leaderboard" target="_blank">
32
+ <img alt="Leaderboard" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Bench-Leaderboard-ffc107?color=ffc107&logoColor=white" height="20" />
33
+ </a>
34
+
35
+ ## Overview
36
+ Video reasoning grounds intelligence in spatiotemporally consistent visual environments that go beyond what text can naturally capture,
37
+ enabling intuitive reasoning over motion, interaction, and causality. Rapid progress in video models has focused primarily on visual quality.
38
+ Systematically studying video reasoning and its scaling behavior suffers from a lack of video reasoning (training) data.
39
+
40
+ To address this gap, we introduce the Very Big Video Reasoning (VBVR) Dataset, an unprecedentedly large-scale resource spanning 200 curated reasoning tasks
41
+ and over one million video clips—approximately three orders of magnitude larger than existing datasets. We further present VBVR-Bench,
42
+ a verifiable evaluation framework that moves beyond model-based judging by incorporating rule-based, human-aligned scorers,
43
+ enabling reproducible and interpretable diagnosis of video reasoning capabilities.
44
+
45
+ Leveraging the VBVR suite, we conduct one of the first large-scale scaling studies of video reasoning and observe early signs of emergent generalization
46
+ to unseen reasoning tasks. **Together, VBVR lays a foundation for the next stage of research in generalizable video reasoning.**
47
+
48
+ The model was presented in the paper [A Very Big Video Reasoning Suite](https://huggingface.co/papers/2602.20159).
49
+
50
+ ## Models Zoo
51
+
52
+ | Model | Base Architecture | Other Remarks |
53
+ |-------|-------------------|---------------|
54
+ | [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) | Wan2.1-I2V-14B-720P | Diffusers format |
55
+ | [VBVR-Wan2.2](https://huggingface.co/Video-Reason/VBVR-Wan2.2) | Wan2.2-I2V-A14B | Diffusers format |
56
+ | [VBVR-Wan2.1-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) | Wan2.1-I2V-14B-720P | DiffSynth LoRA format |
57
+ | [VBVR-Wan2.2-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.2-diffsynth) | Wan2.2-I2V-A14B | DiffSynth LoRA format |
58
+ | [VBVR-LTX2.3-diffsynth](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) | LTX-Video-2.3 | DiffSynth LoRA format |
59
+
60
+ ## Release Information
61
+ VBVR-Wan2.1 is trained from Wan2.1-I2V-14B-720P without architectural modifications, as the goal of VBVR is to *investigate data scaling behavior* and provide *strong baseline models* for the video reasoning research community. Leveraging the VBVR-Dataset, which constitutes one of the largest video reasoning datasets to date, the VBVR model family achieved highest scores on VBVR-Bench.
62
+
63
+ In this release, we present
64
+ [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) (Diffusers format),
65
+ [**VBVR-Wan2.1-diffsynth**](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) (DiffSynth LoRA format), and
66
+ [**VBVR-LTX2.3-diffsynth**](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) (DiffSynth LoRA format; Diffusers does not yet support LTX-Video-2.3, so only the DiffSynth LoRA format is released for this model).
67
+
68
+ <table>
69
+ <tr>
70
+ <th>Model</th>
71
+ <th>Overall</th>
72
+ <th>ID</th>
73
+ <th>ID-Abst.</th>
74
+ <th>ID-Know.</th>
75
+ <th>ID-Perc.</th>
76
+ <th>ID-Spat.</th>
77
+ <th>ID-Trans.</th>
78
+ <th>OOD</th>
79
+ <th>OOD-Abst.</th>
80
+ <th>OOD-Know.</th>
81
+ <th>OOD-Perc.</th>
82
+ <th>OOD-Spat.</th>
83
+ <th>OOD-Trans.</th>
84
+ </tr>
85
+ <tbody>
86
+ <tr>
87
+ <td><strong>Human</strong></td>
88
+ <td>0.974</td><td>0.960</td><td>0.919</td><td>0.956</td><td>1.00</td><td>0.95</td><td>1.00</td>
89
+ <td>0.988</td><td>1.00</td><td>1.00</td><td>0.990</td><td>1.00</td><td>0.970</td>
90
+ </tr>
91
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
92
+ <td colspan="14"><em>Open-source Models</em></td>
93
+ </tr>
94
+ <tr>
95
+ <td>CogVideoX1.5-5B-I2V</td>
96
+ <td>0.273</td><td>0.283</td><td>0.241</td><td>0.328</td><td>0.257</td><td>0.328</td><td>0.305</td>
97
+ <td>0.262</td><td><u>0.281</u></td><td>0.235</td><td>0.250</td><td><strong>0.254</strong></td><td>0.282</td>
98
+ </tr>
99
+ <tr>
100
+ <td>HunyuanVideo-I2V</td>
101
+ <td>0.273</td><td>0.280</td><td>0.207</td><td>0.357</td><td>0.293</td><td>0.280</td><td><u>0.316</u></td>
102
+ <td>0.265</td><td>0.175</td><td><strong>0.369</strong></td><td>0.290</td><td><u>0.253</u></td><td>0.250</td>
103
+ </tr>
104
+ <tr>
105
+ <td><strong>Wan2.2-I2V-A14B</strong></td>
106
+ <td><strong>0.371</strong></td><td><strong>0.412</strong></td><td><strong>0.430</strong></td>
107
+ <td><strong>0.382</strong></td><td><strong>0.415</strong></td><td><strong>0.404</strong></td>
108
+ <td><strong>0.419</strong></td><td><strong>0.329</strong></td>
109
+ <td><strong>0.405</strong></td><td>0.308</td><td><strong>0.343</strong></td>
110
+ <td>0.236</td><td><u>0.307</u></td>
111
+ </tr>
112
+ <tr>
113
+ <td><u>LTX-2</u></td>
114
+ <td><u>0.313</u></td><td><u>0.329</u></td><td><u>0.316</u></td>
115
+ <td><u>0.362</u></td><td><u>0.326</u></td><td><u>0.340</u></td>
116
+ <td>0.306</td><td><u>0.297</u></td>
117
+ <td>0.244</td><td><u>0.337</u></td><td><u>0.317</u></td>
118
+ <td>0.231</td><td><strong>0.311</strong></td>
119
+ </tr>
120
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
121
+ <td colspan="14"><em>Proprietary Models</em></td>
122
+ </tr>
123
+ <tr>
124
+ <td><u>Seedance 2.0</u></td>
125
+ <td><u>0.544</u></td><td><strong>0.570</strong></td><td>0.593</td><td><u>0.498</u></td><td><strong>0.618</strong></td><td><u>0.514</u></td><td><strong>0.602</strong></td>
126
+ <td><u>0.517</u></td><td><strong>0.643</strong></td><td>0.398</td><td><u>0.492</u></td><td>0.427</td><td><strong>0.556</strong></td>
127
+ </tr>
128
+ <tr>
129
+ <td>Runway Gen-4 Turbo</td>
130
+ <td>0.403</td><td>0.392</td><td>0.396</td><td>0.409</td><td>0.429</td><td>0.341</td><td>0.363</td>
131
+ <td>0.414</td><td>0.515</td><td><u>0.429</u></td><td>0.419</td><td>0.327</td><td>0.373</td>
132
+ </tr>
133
+ <tr>
134
+ <td><strong>Sora 2</strong></td>
135
+ <td><strong>0.546</strong></td><td><u>0.569</u></td><td><u>0.602</u></td>
136
+ <td>0.477</td><td><u>0.581</u></td><td><strong>0.572</strong></td>
137
+ <td><u>0.597</u></td><td><strong>0.523</strong></td>
138
+ <td><u>0.546</u></td><td><strong>0.472</strong></td><td><strong>0.525</strong></td>
139
+ <td><strong>0.462</strong></td><td><u>0.546</u></td>
140
+ </tr>
141
+ <tr>
142
+ <td>Kling 2.6</td>
143
+ <td>0.369</td><td>0.408</td><td>0.465</td><td>0.323</td><td>0.375</td><td>0.347</td><td>0.519</td>
144
+ <td>0.330</td><td>0.528</td><td>0.135</td><td>0.272</td><td>0.356</td><td>0.359</td>
145
+ </tr>
146
+ <tr>
147
+ <td>Veo 3.1</td>
148
+ <td>0.480</td><td>0.531</td><td><strong>0.611</strong></td>
149
+ <td><strong>0.503</strong></td><td>0.520</td><td>0.444</td>
150
+ <td>0.510</td><td>0.429</td>
151
+ <td><u>0.577</u></td><td>0.277</td><td>0.420</td>
152
+ <td><u>0.441</u></td><td>0.404</td>
153
+ </tr>
154
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
155
+ <td colspan="14"><em>Data Scaling Strong Baseline</em></td>
156
+ </tr>
157
+ <tr>
158
+ <td><strong>VBVR-LTX2.3</strong></td>
159
+ <td>0.516</td><td>0.580</td><td>0.608</td><td>0.631</td><td>0.529</td><td>0.454</td><td>0.680</td>
160
+ <td>0.453</td><td>0.608</td><td>0.577</td><td><u>0.409</u></td><td>0.414</td><td><u>0.388</u></td>
161
+ </tr>
162
+ <tr>
163
+ <td><strong>VBVR-Wan2.1</strong></td>
164
+ <td><u>0.592</u></td><td><u>0.724</u></td><td><u>0.705</u></td><td><u>0.710</u></td><td><u>0.727</u></td><td><u>0.719</u></td><td><u>0.784</u></td>
165
+ <td><u>0.461</u></td><td><u>0.674</u></td><td><strong>0.592</strong></td><td>0.387</td><td><u>0.461</u></td><td>0.387</td>
166
+ </tr>
167
+ <tr>
168
+ <td><strong>VBVR-Wan2.2</strong></td>
169
+ <td><strong>0.685</strong></td><td><strong>0.760</strong></td><td><strong>0.724</strong></td>
170
+ <td><strong>0.750</strong></td><td><strong>0.782</strong></td><td><strong>0.745</strong></td>
171
+ <td><strong>0.833</strong></td><td><strong>0.610</strong></td>
172
+ <td><strong>0.768</strong></td><td><u>0.572</u></td><td><strong>0.547</strong></td>
173
+ <td><strong>0.618</strong></td><td><strong>0.615</strong></td>
174
+ </tr>
175
+ </tbody>
176
+ </table>
177
+
178
+ ## QuickStart
179
+
180
+ ### Installation
181
+
182
+ We recommend using [uv](https://docs.astral.sh/uv/) to manage the environment.
183
+
184
+ > uv installation guide: <https://docs.astral.sh/uv/getting-started/installation/#installing-uv>
185
+
186
+ ```bash
187
+ pip install torch>=2.4.0 torchvision>=0.19.0 transformers Pillow huggingface_hub[cli]
188
+ uv pip install git+https://github.com/huggingface/diffusers
189
+ ```
190
+
191
+ ### Example Code
192
+
193
+ ```bash
194
+ huggingface-cli download Video-Reason/VBVR-Wan2.1 --local-dir ./VBVR-Wan2.1
195
+ python example.py \
196
+ --model_path ./VBVR-Wan2.1
197
+ ```
198
+
199
+ ## Citation
200
+
201
+ ```bibtex
202
+ @article{vbvr2026,
203
+ title = {A Very Big Video Reasoning Suite},
204
+ author = {Wang, Maijunxian and Wang, Ruisi and Lin, Juyi and Ji, Ran and
205
+ Wiedemer, Thadd{\"a}us and Gao, Qingying and Luo, Dezhi and
206
+ Qian, Yaoyao and Huang, Lianyu and Hong, Zelong and Ge, Jiahui and
207
+ Ma, Qianli and He, Hang and Zhou, Yifan and Guo, Lingzi and
208
+ Mei, Lantao and Li, Jiachen and Xing, Hanwen and Zhao, Tianqi and
209
+ Yu, Fengyuan and Xiao, Weihang and Jiao, Yizheng and
210
+ Hou, Jianheng and Zhang, Danyang and Xu, Pengcheng and
211
+ Zhong, Boyang and Zhao, Zehong and Fang, Gaoyun and Kitaoka, John and
212
+ Xu, Yile and Xu, Hua bureau and Blacutt, Kenton and Nguyen, Tin and
213
+ Song, Siyuan and Sun, Haoran and Wen, Shaoyue and He, Linyang and
214
+ Wang, Runming and Wang, Yanzhi and Yang, Mengyue and Ma, Ziqiao and
215
+ Milli{\`e}re, Rapha{\"e}l and Shi, Freda and Vasconcelos, Nuno and
216
+ Khashabi, Daniel and Yuille, Alan and Du, Yilun and Liu, Ziming and
217
+ Lin, Dahua and Liu, Ziwei and Kumar, Vikash and Li, Yijiang and
218
+ Yang, Lei and Cai, Zhongang and Deng, Hokin},
219
+ journal = {arXiv preprint arXiv:2602.20159},
220
+ year = {2026},
221
+ url = {https://arxiv.org/abs/2602.20159}
222
+ }
223
+ ```
lora.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53e681a110bc0a194a5d01d72dcb448cdfb1cf00249b6f27b92cd262009a16ea
3
+ size 428132264