Update run_wan_I2V_CP+FB.py
Browse files- run_wan_I2V_CP+FB.py +2 -1
run_wan_I2V_CP+FB.py
CHANGED
|
@@ -47,7 +47,8 @@ end_loading = time.time()
|
|
| 47 |
from para_attn.first_block_cache.diffusers_adapters import apply_cache_on_pipe
|
| 48 |
apply_cache_on_pipe(pipe , residual_diff_threshold=0.1)
|
| 49 |
|
| 50 |
-
pipe.enable_model_cpu_offload()
|
|
|
|
| 51 |
|
| 52 |
# torch._inductor.config.reorder_for_compute_comm_overlap = True
|
| 53 |
# pipe.transformer = torch.compile(pipe.transformer, mode="max-autotune-no-cudagraphs")
|
|
|
|
| 47 |
from para_attn.first_block_cache.diffusers_adapters import apply_cache_on_pipe
|
| 48 |
apply_cache_on_pipe(pipe , residual_diff_threshold=0.1)
|
| 49 |
|
| 50 |
+
# pipe.enable_model_cpu_offload(gpu_id=dist.get_rank())
|
| 51 |
+
pipe.enable_vae_tiling()
|
| 52 |
|
| 53 |
# torch._inductor.config.reorder_for_compute_comm_overlap = True
|
| 54 |
# pipe.transformer = torch.compile(pipe.transformer, mode="max-autotune-no-cudagraphs")
|