Image-to-Video
Diffusers
wruisi commited on
Commit
0f43af4
·
1 Parent(s): d9e2a9f

Upload model

Browse files
Files changed (2) hide show
  1. README.md +208 -3
  2. lora.safetensors +3 -0
README.md CHANGED
@@ -1,3 +1,208 @@
1
- ---
2
- license: cc
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Wan-AI/Wan2.1-I2V-14B-720P-Diffusers
4
+ library_name: diffusers
5
+ license: apache-2.0
6
+ pipeline_tag: image-to-video
7
+ ---
8
+ # VBVR: A Very Big Video Reasoning Suite
9
+
10
+ <a href="https://video-reason.com" target="_blank">
11
+ <img alt="Project Page" src="https://img.shields.io/badge/Project%20-%20Homepage-4285F4" height="20" />
12
+ </a>
13
+ <a href="https://github.com/Video-Reason/VBVR-EvalKit" target="_blank">
14
+ <img alt="Code" src="https://img.shields.io/badge/Evaluation_code-VBVR_Bench-100000?style=flat-square&logo=github&logoColor=white" height="20" />
15
+ </a>
16
+ <a href="https://github.com/Video-Reason/VBVR-Wan2.2" target="_blank">
17
+ <img alt="Code" src="https://img.shields.io/badge/Training_code-VBVR_Wan2.2-100000?style=flat-square&logo=github&logoColor=white" height="20" />
18
+ </a>
19
+ <a href="https://github.com/Video-Reason/VBVR-DataFactory" target="_blank">
20
+ <img alt="Code" src="https://img.shields.io/badge/Data_code-VBVR_DataFactory-100000?style=flat-square&logo=github&logoColor=white" height="20" />
21
+ </a>
22
+ <a href="https://huggingface.co/papers/2602.20159" target="_blank">
23
+ <img alt="arXiv" src="https://img.shields.io/badge/arXiv-VBVR-red?logo=arxiv" height="20" />
24
+ </a>
25
+ <a href="https://huggingface.co/datasets/Video-Reason/VBVR-Dataset" target="_blank">
26
+ <img alt="Dataset" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Dataset-Data-ffc107?color=ffc107&logoColor=white" height="20" />
27
+ </a>
28
+ <a href="https://huggingface.co/datasets/Video-Reason/VBVR-Bench-Data" target="_blank">
29
+ <img alt="Bench Data" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Bench-Data-ffc107?color=ffc107&logoColor=white" height="20" />
30
+ </a>
31
+ <a href="https://huggingface.co/spaces/Video-Reason/VBVR-Bench-Leaderboard" target="_blank">
32
+ <img alt="Leaderboard" src="https://img.shields.io/badge/%F0%9F%A4%97%20_VBVR_Bench-Leaderboard-ffc107?color=ffc107&logoColor=white" height="20" />
33
+ </a>
34
+
35
+ ## Overview
36
+ Video reasoning grounds intelligence in spatiotemporally consistent visual environments that go beyond what text can naturally capture,
37
+ enabling intuitive reasoning over motion, interaction, and causality. Rapid progress in video models has focused primarily on visual quality.
38
+ Systematically studying video reasoning and its scaling behavior suffers from a lack of video reasoning (training) data.
39
+
40
+ To address this gap, we introduce the Very Big Video Reasoning (VBVR) Dataset, an unprecedentedly large-scale resource spanning 200 curated reasoning tasks
41
+ and over one million video clips—approximately three orders of magnitude larger than existing datasets. We further present VBVR-Bench,
42
+ a verifiable evaluation framework that moves beyond model-based judging by incorporating rule-based, human-aligned scorers,
43
+ enabling reproducible and interpretable diagnosis of video reasoning capabilities.
44
+
45
+ Leveraging the VBVR suite, we conduct one of the first large-scale scaling studies of video reasoning and observe early signs of emergent generalization
46
+ to unseen reasoning tasks. **Together, VBVR lays a foundation for the next stage of research in generalizable video reasoning.**
47
+
48
+ The model was presented in the paper [A Very Big Video Reasoning Suite](https://huggingface.co/papers/2602.20159).
49
+
50
+ ## Models Zoo
51
+
52
+ | Model | Base Architecture | Other Remarks |
53
+ |-------|-------------------|---------------|
54
+ | [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) | Wan2.1-I2V-14B-720P | Diffusers format |
55
+ | [VBVR-Wan2.2](https://huggingface.co/Video-Reason/VBVR-Wan2.2) | Wan2.2-I2V-A14B | Diffusers format |
56
+ | [VBVR-Wan2.1-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) | Wan2.1-I2V-14B-720P | DiffSynth LoRA format |
57
+ | [VBVR-Wan2.2-diffsynth](https://huggingface.co/Video-Reason/VBVR-Wan2.2-diffsynth) | Wan2.2-I2V-A14B | DiffSynth LoRA format |
58
+ | [VBVR-LTX2.3-diffsynth](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) | LTX-Video-2.3 | DiffSynth LoRA format |
59
+
60
+ ## Release Information
61
+ VBVR-Wan2.1 is trained from Wan2.1-I2V-14B-720P without architectural modifications, as the goal of VBVR is to *investigate data scaling behavior* and provide *strong baseline models* for the video reasoning research community. Leveraging the VBVR-Dataset, which constitutes one of the largest video reasoning datasets to date, the VBVR model family achieved highest scores on VBVR-Bench.
62
+
63
+ In this release, we present
64
+ [**VBVR-Wan2.1**](https://huggingface.co/Video-Reason/VBVR-Wan2.1) (Diffusers format),
65
+ [**VBVR-Wan2.1-diffsynth**](https://huggingface.co/Video-Reason/VBVR-Wan2.1-diffsynth) (DiffSynth LoRA format), and
66
+ [**VBVR-LTX2.3-diffsynth**](https://huggingface.co/Video-Reason/VBVR-LTX2.3-diffsynth) (DiffSynth LoRA format; Diffusers does not yet support LTX-Video-2.3, so only the DiffSynth LoRA format is released for this model).
67
+
68
+ <table>
69
+ <tr>
70
+ <th>Model</th>
71
+ <th>Overall</th>
72
+ <th>ID</th>
73
+ <th>ID-Abst.</th>
74
+ <th>ID-Know.</th>
75
+ <th>ID-Perc.</th>
76
+ <th>ID-Spat.</th>
77
+ <th>ID-Trans.</th>
78
+ <th>OOD</th>
79
+ <th>OOD-Abst.</th>
80
+ <th>OOD-Know.</th>
81
+ <th>OOD-Perc.</th>
82
+ <th>OOD-Spat.</th>
83
+ <th>OOD-Trans.</th>
84
+ </tr>
85
+ <tbody>
86
+ <tr>
87
+ <td><strong>Human</strong></td>
88
+ <td>0.974</td><td>0.960</td><td>0.919</td><td>0.956</td><td>1.00</td><td>0.95</td><td>1.00</td>
89
+ <td>0.988</td><td>1.00</td><td>1.00</td><td>0.990</td><td>1.00</td><td>0.970</td>
90
+ </tr>
91
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
92
+ <td colspan="14"><em>Open-source Models</em></td>
93
+ </tr>
94
+ <tr>
95
+ <td>CogVideoX1.5-5B-I2V</td>
96
+ <td>0.273</td><td>0.283</td><td>0.241</td><td>0.328</td><td>0.257</td><td>0.328</td><td>0.305</td>
97
+ <td>0.262</td><td><u>0.281</u></td><td>0.235</td><td>0.250</td><td><strong>0.254</strong></td><td>0.282</td>
98
+ </tr>
99
+ <tr>
100
+ <td>HunyuanVideo-I2V</td>
101
+ <td>0.273</td><td>0.280</td><td>0.207</td><td>0.357</td><td>0.293</td><td>0.280</td><td><u>0.316</u></td>
102
+ <td>0.265</td><td>0.175</td><td><strong>0.369</strong></td><td>0.290</td><td><u>0.253</u></td><td>0.250</td>
103
+ </tr>
104
+ <tr>
105
+ <td><strong>Wan2.2-I2V-A14B</strong></td>
106
+ <td><strong>0.371</strong></td><td><strong>0.412</strong></td><td><strong>0.430</strong></td>
107
+ <td><strong>0.382</strong></td><td><strong>0.415</strong></td><td><strong>0.404</strong></td>
108
+ <td><strong>0.419</strong></td><td><strong>0.329</strong></td>
109
+ <td><strong>0.405</strong></td><td>0.308</td><td><strong>0.343</strong></td>
110
+ <td>0.236</td><td><u>0.307</u></td>
111
+ </tr>
112
+ <tr>
113
+ <td><u>LTX-2</u></td>
114
+ <td><u>0.313</u></td><td><u>0.329</u></td><td><u>0.316</u></td>
115
+ <td><u>0.362</u></td><td><u>0.326</u></td><td><u>0.340</u></td>
116
+ <td>0.306</td><td><u>0.297</u></td>
117
+ <td>0.244</td><td><u>0.337</u></td><td><u>0.317</u></td>
118
+ <td>0.231</td><td><strong>0.311</strong></td>
119
+ </tr>
120
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
121
+ <td colspan="14"><em>Proprietary Models</em></td>
122
+ </tr>
123
+ <tr>
124
+ <td><u>Seedance 2.0</u></td>
125
+ <td><u>0.544</u></td><td><strong>0.570</strong></td><td>0.593</td><td><u>0.498</u></td><td><strong>0.618</strong></td><td><u>0.514</u></td><td><strong>0.602</strong></td>
126
+ <td><u>0.517</u></td><td><strong>0.643</strong></td><td>0.398</td><td><u>0.492</u></td><td>0.427</td><td><strong>0.556</strong></td>
127
+ </tr>
128
+ <tr>
129
+ <td>Runway Gen-4 Turbo</td>
130
+ <td>0.403</td><td>0.392</td><td>0.396</td><td>0.409</td><td>0.429</td><td>0.341</td><td>0.363</td>
131
+ <td>0.414</td><td>0.515</td><td><u>0.429</u></td><td>0.419</td><td>0.327</td><td>0.373</td>
132
+ </tr>
133
+ <tr>
134
+ <td><strong>Sora 2</strong></td>
135
+ <td><strong>0.546</strong></td><td><u>0.569</u></td><td><u>0.602</u></td>
136
+ <td>0.477</td><td><u>0.581</u></td><td><strong>0.572</strong></td>
137
+ <td><u>0.597</u></td><td><strong>0.523</strong></td>
138
+ <td><u>0.546</u></td><td><strong>0.472</strong></td><td><strong>0.525</strong></td>
139
+ <td><strong>0.462</strong></td><td><u>0.546</u></td>
140
+ </tr>
141
+ <tr>
142
+ <td>Kling 2.6</td>
143
+ <td>0.369</td><td>0.408</td><td>0.465</td><td>0.323</td><td>0.375</td><td>0.347</td><td>0.519</td>
144
+ <td>0.330</td><td>0.528</td><td>0.135</td><td>0.272</td><td>0.356</td><td>0.359</td>
145
+ </tr>
146
+ <tr>
147
+ <td>Veo 3.1</td>
148
+ <td>0.480</td><td>0.531</td><td><strong>0.611</strong></td>
149
+ <td><strong>0.503</strong></td><td>0.520</td><td>0.444</td>
150
+ <td>0.510</td><td>0.429</td>
151
+ <td><u>0.577</u></td><td>0.277</td><td>0.420</td>
152
+ <td><u>0.441</u></td><td>0.404</td>
153
+ </tr>
154
+ <tr style="background:#F2F0EF;font-weight:700;text-align:center;">
155
+ <td colspan="14"><em>Data Scaling Strong Baseline</em></td>
156
+ </tr>
157
+ <tr>
158
+ <td><strong>VBVR-LTX2.3</strong></td>
159
+ <td>0.516</td><td>0.580</td><td>0.608</td><td>0.631</td><td>0.529</td><td>0.454</td><td>0.680</td>
160
+ <td>0.453</td><td>0.608</td><td>0.577</td><td><u>0.409</u></td><td>0.414</td><td><u>0.388</u></td>
161
+ </tr>
162
+ <tr>
163
+ <td><strong>VBVR-Wan2.1</strong></td>
164
+ <td><u>0.592</u></td><td><u>0.724</u></td><td><u>0.705</u></td><td><u>0.710</u></td><td><u>0.727</u></td><td><u>0.719</u></td><td><u>0.784</u></td>
165
+ <td><u>0.461</u></td><td><u>0.674</u></td><td><strong>0.592</strong></td><td>0.387</td><td><u>0.461</u></td><td>0.387</td>
166
+ </tr>
167
+ <tr>
168
+ <td><strong>VBVR-Wan2.2</strong></td>
169
+ <td><strong>0.685</strong></td><td><strong>0.760</strong></td><td><strong>0.724</strong></td>
170
+ <td><strong>0.750</strong></td><td><strong>0.782</strong></td><td><strong>0.745</strong></td>
171
+ <td><strong>0.833</strong></td><td><strong>0.610</strong></td>
172
+ <td><strong>0.768</strong></td><td><u>0.572</u></td><td><strong>0.547</strong></td>
173
+ <td><strong>0.618</strong></td><td><strong>0.615</strong></td>
174
+ </tr>
175
+ </tbody>
176
+ </table>
177
+
178
+ ## QuickStart
179
+
180
+ ### Inference
181
+ For running inference, please refer to the [**official guide**](https://github.com/Video-Reason/VBVR-Wan2.2?tab=readme-ov-file#wan21-inference) in the VBVR-Wan2.2 GitHub repository.
182
+ This repository contains the latest instructions, configurations, and examples for performing inference with the VBVR family models.
183
+
184
+ ## Citation
185
+
186
+ ```bibtex
187
+ @article{vbvr2026,
188
+ title = {A Very Big Video Reasoning Suite},
189
+ author = {Wang, Maijunxian and Wang, Ruisi and Lin, Juyi and Ji, Ran and
190
+ Wiedemer, Thadd{\"a}us and Gao, Qingying and Luo, Dezhi and
191
+ Qian, Yaoyao and Huang, Lianyu and Hong, Zelong and Ge, Jiahui and
192
+ Ma, Qianli and He, Hang and Zhou, Yifan and Guo, Lingzi and
193
+ Mei, Lantao and Li, Jiachen and Xing, Hanwen and Zhao, Tianqi and
194
+ Yu, Fengyuan and Xiao, Weihang and Jiao, Yizheng and
195
+ Hou, Jianheng and Zhang, Danyang and Xu, Pengcheng and
196
+ Zhong, Boyang and Zhao, Zehong and Fang, Gaoyun and Kitaoka, John and
197
+ Xu, Yile and Xu, Hua bureau and Blacutt, Kenton and Nguyen, Tin and
198
+ Song, Siyuan and Sun, Haoran and Wen, Shaoyue and He, Linyang and
199
+ Wang, Runming and Wang, Yanzhi and Yang, Mengyue and Ma, Ziqiao and
200
+ Milli{\`e}re, Rapha{\"e}l and Shi, Freda and Vasconcelos, Nuno and
201
+ Khashabi, Daniel and Yuille, Alan and Du, Yilun and Liu, Ziming and
202
+ Lin, Dahua and Liu, Ziwei and Kumar, Vikash and Li, Yijiang and
203
+ Yang, Lei and Cai, Zhongang and Deng, Hokin},
204
+ journal = {arXiv preprint arXiv:2602.20159},
205
+ year = {2026},
206
+ url = {https://arxiv.org/abs/2602.20159}
207
+ }
208
+ ```
lora.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8345fad02f752c6b2f5de4c44f9c73d5bc6e8cf5d26e3224e16be5b064229a1
3
+ size 306801576