| """ |
| Convert Diffusers-format FLUX model to ComfyUI-compatible checkpoint. |
| This handles the proper folder structure and key naming. |
| """ |
|
|
| from safetensors.torch import save_file, load_file |
| import os |
| import json |
| from pathlib import Path |
|
|
| def convert_diffusers_to_comfyui( |
| diffusers_folder, |
| output_path, |
| fp16=False |
| ): |
| """ |
| Convert a Diffusers FLUX model folder to a single ComfyUI checkpoint. |
| |
| Args: |
| diffusers_folder: Path to folder containing model_index.json |
| output_path: Output path for the merged .safetensors file |
| fp16: If True, convert to float16 to save space |
| """ |
| |
| diffusers_folder = Path(diffusers_folder) |
| |
| |
| model_index = diffusers_folder / "model_index.json" |
| if not model_index.exists(): |
| raise ValueError(f"Not a Diffusers model folder. Missing: {model_index}") |
| |
| with open(model_index) as f: |
| config = json.load(f) |
| |
| print("=" * 80) |
| print("DIFFUSERS TO COMFYUI CONVERTER") |
| print("=" * 80) |
| print(f"\nModel: {config.get('_name_or_path', 'Unknown')}") |
| print(f"Format: {config.get('_class_name', 'Unknown')}") |
| |
| merged_state = {} |
| |
| |
| |
| |
| print("\n" + "=" * 80) |
| print("Loading Transformer...") |
| print("=" * 80) |
| |
| transformer_path = diffusers_folder / "transformer" |
| transformer_file = None |
| |
| |
| for file in transformer_path.glob("*.safetensors"): |
| transformer_file = file |
| break |
| |
| if not transformer_file: |
| raise ValueError(f"No safetensors file found in {transformer_path}") |
| |
| print(f"Found: {transformer_file.name}") |
| transformer_state = load_file(str(transformer_file)) |
| print(f"Loaded {len(transformer_state)} transformer parameters") |
| |
| |
| for key, value in transformer_state.items(): |
| if fp16 and value.dtype.is_floating_point: |
| value = value.half() |
| merged_state[key] = value |
| |
| |
| |
| |
| print("\n" + "=" * 80) |
| print("Loading VAE...") |
| print("=" * 80) |
| |
| vae_path = diffusers_folder / "vae" |
| vae_file = None |
| |
| for file in vae_path.glob("*.safetensors"): |
| vae_file = file |
| break |
| |
| if not vae_file: |
| print("⚠️ No VAE file found, skipping...") |
| else: |
| print(f"Found: {vae_file.name}") |
| vae_state = load_file(str(vae_file)) |
| print(f"Loaded {len(vae_state)} VAE parameters") |
| |
| |
| for key, value in vae_state.items(): |
| if fp16 and value.dtype.is_floating_point: |
| value = value.half() |
| |
| merged_state[key] = value |
| |
| |
| |
| |
| print("\n" + "=" * 80) |
| print("Loading Text Encoders...") |
| print("=" * 80) |
| |
| |
| clip_path = diffusers_folder / "text_encoder" |
| if clip_path.exists(): |
| clip_file = None |
| for file in clip_path.glob("*.safetensors"): |
| clip_file = file |
| break |
| |
| if clip_file: |
| print(f"Found CLIP: {clip_file.name}") |
| clip_state = load_file(str(clip_file)) |
| print(f"Loaded {len(clip_state)} CLIP parameters") |
| |
| for key, value in clip_state.items(): |
| if fp16 and value.dtype.is_floating_point: |
| value = value.half() |
| |
| merged_state[key] = value |
| else: |
| print("⚠️ No CLIP file found") |
| |
| |
| t5_path = diffusers_folder / "text_encoder_2" |
| if t5_path.exists(): |
| t5_file = None |
| for file in t5_path.glob("*.safetensors"): |
| t5_file = file |
| break |
| |
| if t5_file: |
| print(f"Found T5: {t5_file.name}") |
| print("⚠️ Loading T5 (this may take a while, it's large)...") |
| t5_state = load_file(str(t5_file)) |
| print(f"Loaded {len(t5_state)} T5 parameters") |
| |
| for key, value in t5_state.items(): |
| if fp16 and value.dtype.is_floating_point: |
| value = value.half() |
| merged_state[key] = value |
| else: |
| print("⚠️ No T5 file found") |
| |
| |
| |
| |
| print("\n" + "=" * 80) |
| print("Saving merged checkpoint...") |
| print("=" * 80) |
| |
| print(f"Total parameters: {len(merged_state):,}") |
| print(f"Output: {output_path}") |
| |
| save_file(merged_state, output_path) |
| |
| size_gb = os.path.getsize(output_path) / (1024**3) |
| print(f"\n✅ Conversion complete!") |
| print(f"File size: {size_gb:.2f} GB") |
| |
| |
| print("\n" + "=" * 80) |
| print("Key Structure in Merged File") |
| print("=" * 80) |
| |
| sample_keys = list(merged_state.keys())[:10] |
| print("\nFirst 10 keys:") |
| for key in sample_keys: |
| print(f" {key}") |
| |
| return output_path |
|
|
|
|
| def convert_with_working_template( |
| diffusers_folder, |
| working_checkpoint, |
| output_path, |
| replace_transformer_only=True |
| ): |
| """ |
| Use a working checkpoint as template, replacing components from Diffusers model. |
| This ensures key naming matches what ComfyUI expects. |
| |
| Args: |
| diffusers_folder: Path to Diffusers model folder |
| working_checkpoint: Path to a working ComfyUI checkpoint |
| output_path: Output path for merged checkpoint |
| replace_transformer_only: If True, only replace transformer, keep VAE/encoders from template |
| """ |
| |
| print("=" * 80) |
| print("TEMPLATE-BASED CONVERSION") |
| print("=" * 80) |
| |
| |
| print("\nLoading template checkpoint...") |
| template_state = load_file(working_checkpoint) |
| print(f"Template has {len(template_state)} parameters") |
| |
| |
| template_keys = set(template_state.keys()) |
| transformer_keys = {k for k in template_keys if 'transformer' in k or 'double_blocks' in k or 'single_blocks' in k} |
| vae_keys = {k for k in template_keys if 'vae' in k.lower() or 'first_stage' in k} |
| text_encoder_keys = {k for k in template_keys if 'text_encoder' in k or 'clip' in k.lower()} |
| |
| print(f"\nTemplate structure:") |
| print(f" Transformer keys: {len(transformer_keys)}") |
| print(f" VAE keys: {len(vae_keys)}") |
| print(f" Text encoder keys: {len(text_encoder_keys)}") |
| |
| |
| diffusers_folder = Path(diffusers_folder) |
| transformer_path = diffusers_folder / "transformer" |
| |
| transformer_file = next(transformer_path.glob("*.safetensors")) |
| print(f"\nLoading new transformer from: {transformer_file.name}") |
| new_transformer = load_file(str(transformer_file)) |
| |
| |
| print("\nReplacing transformer weights...") |
| merged_state = dict(template_state) |
| |
| |
| replaced = 0 |
| for key in transformer_keys: |
| if key in new_transformer: |
| merged_state[key] = new_transformer[key] |
| replaced += 1 |
| |
| print(f"Replaced {replaced} transformer parameters") |
| |
| if not replace_transformer_only: |
| print("\n⚠️ Also replacing VAE and text encoders...") |
| |
| vae_file = next((diffusers_folder / "vae").glob("*.safetensors"), None) |
| if vae_file: |
| vae_state = load_file(str(vae_file)) |
| for key in vae_keys: |
| if key in vae_state: |
| merged_state[key] = vae_state[key] |
| |
| |
| |
| |
| print(f"\nSaving to {output_path}...") |
| save_file(merged_state, output_path) |
| |
| size_gb = os.path.getsize(output_path) / (1024**3) |
| print(f"✅ Done! File size: {size_gb:.2f} GB") |
|
|
|
|
| |
| if __name__ == "__main__": |
| |
| |
| |
| |
| |
| |
| |
| |
| convert_with_working_template( |
| diffusers_folder="../", |
| working_checkpoint="../quantized/svdq-fp4_r32-flux.1-depth-dev.safetensors", |
| output_path="svdq-fp4_r32-flux.1-depth-dev_ComfyMerged.safetensors", |
| replace_transformer_only=True |
| ) |
|
|