james16 commited on
Commit
338ed35
Β·
verified Β·
1 Parent(s): 758af8f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -3
README.md CHANGED
@@ -1,3 +1,54 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: diffusers
4
+ tags:
5
+ - computer-vision
6
+ - video-editing
7
+ - video-to-video
8
+ - diffusion
9
+ - flow-matching
10
+ - cvpr2026
11
+ ---
12
+
13
+ # [CVPR 2026] PropFly: Learning to Propagate via On-the-Fly Supervision from Pre-trained Video Diffusion Models
14
+
15
+ <div align="left">
16
+ <a href="https://kaist-viclab.github.io/PropFly_site/"><img src="https://img.shields.io/badge/Project-Page-blue" alt="Project Page"></a>
17
+ <a href="https://arxiv.org/abs/2602.20583"><img src="https://img.shields.io/badge/arXiv-2602.20583-b31b1b.svg" alt="arXiv"></a>
18
+ <a href="https://github.com/pmjames16/PropFly"><img src="https://img.shields.io/badge/GitHub-Code-black?logo=github" alt="GitHub"></a>
19
+ </div>
20
+
21
+ Official model weights for **PropFly**.
22
+
23
+ PropFly is a novel training pipeline for propagation-based video editing that eliminates the need for large-scale, paired (source and edited) video datasets. Instead, it leverages on-the-fly supervision from pre-trained Video Diffusion Models (VDMs).
24
+
25
+
26
+ ## Model Description
27
+
28
+ Propagation-based video editing enables precise user control by propagating a single edited frame into subsequent frames while maintaining the original context. Our proposed method, **PropFly**, achieves this by:
29
+
30
+ 1. **On-the-Fly Supervision:** Utilizing a frozen pre-trained VDM to synthesize structurally aligned yet semantically distinct source (low-CFG) and target (high-CFG) latent pairs on the fly.
31
+ 2. **Guidance-Modulated Flow Matching (GMFM):** Training an adapter to learn propagation by predicting the VDM's high-CFG velocity, conditioned on the source video structure and the edited first frame style via GMFM loss.
32
+
33
+ This approach ensures temporally consistent and dynamic transformations, significantly outperforming state-of-the-art methods on various video editing tasks (evaluated on EditVerseBench and TGVE benchmarks).
34
+
35
+ ## Repository Structure
36
+
37
+ The model weights are stored in the `PropFly-1.3B/` directory.
38
+
39
+ ```text
40
+ β”œβ”€β”€ PropFly-1.3B/
41
+ β”‚ β”œβ”€β”€ diffusion_pytorch_model.bin # Model weights
42
+ β”œβ”€β”€ .gitattributes
43
+ └── README.md
44
+ ```
45
+
46
+ ## Citation
47
+ ```text
48
+ @article{seo2026propfly,
49
+ title={PropFly: Learning to Propagate via On-the-Fly Supervision from Pre-trained Video Diffusion Models},
50
+ author={Seo, Wonyong and Moon, Jaeho and Lee, Jaehyup and Kim, Soo Ye and Kim, Munchurl},
51
+ journal={arXiv preprint arXiv:2602.20583},
52
+ year={2026}
53
+ }
54
+ ```