Bruece commited on
Commit
67110a3
ยท
verified ยท
1 Parent(s): ce9d0c3

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ conceptmix_results.png filter=lfs diff=lfs merge=lfs -text
GenEval2_results.png ADDED
README.md ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: apache-2.0
4
+ library_name: diffusers
5
+ base_model: black-forest-labs/FLUX.1-dev
6
+ tags:
7
+ - flux
8
+ - diffusers
9
+ - lora
10
+ - cmo
11
+ - text-to-image
12
+ pipeline_tag: text-to-image
13
+ ---
14
+
15
+ # FLUX.1-dev-CMO
16
+
17
+ <p align="center">
18
+ ๐Ÿค— <a href="[https://huggingface.co/](https://huggingface.co/)Bruece/FLUX.1-dev-CMO"><b>Hugging Face</b></a> |
19
+ ๐Ÿ“„ <a href="[https://arxiv.org/abs/2603.18528](https://arxiv.org/abs/2603.18528)"><b>arXiv</b></a>
20
+ </p>
21
+
22
+ **๐ŸŒŸ Official LoRA Adapter for [Correlation-Weighted Multi-Reward Optimization for Compositional Generation](https://arxiv.org/abs/2603.18528)**
23
+
24
+ This repository contains the official LoRA adapter for [black-forest-labs/FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev) fine-tuned using **CMO (Correlation-Weighted Multi-Reward Optimization)** to enhance compositional generation capabilities.
25
+
26
+ ## ๐Ÿš€ Usage
27
+
28
+ Below is the code to load and merge the LoRA adapter with the base FLUX.1-dev model.
29
+
30
+ ```python
31
+ import torch
32
+ from diffusers import FluxPipeline
33
+ from peft import PeftModel
34
+
35
+ model_id = "black-forest-labs/FLUX.1-dev"
36
+ lora_ckpt_path = "Bruece/FLUX.1-dev-CMO"
37
+ device = "cuda"
38
+
39
+ pipe = FluxPipeline.from_pretrained(model_id, torch_dtype=torch.bfloat16)
40
+ pipe.transformer = PeftModel.from_pretrained(pipe.transformer, lora_ckpt_path)
41
+ pipe.transformer = pipe.transformer.merge_and_unload()
42
+ pipe = pipe.to(device)
43
+
44
+ prompt = 'a photo of a black kite and a green bear'
45
+ image = pipe(prompt, height=512, width=512, num_inference_steps=40, guidance_scale=4.5).images[0]
46
+ image.save("flux_cmo_lora.png")
47
+ ```
48
+
49
+ ## ๐Ÿ–ผ๏ธ Qualitative Results
50
+
51
+ <details>
52
+ <summary>ConceptMix (<a href="[https://arxiv.org/abs/2408.14339](https://arxiv.org/abs/2408.14339)">Link</a>)</summary>
53
+ <br>
54
+ <img src="./conceptmix_results.png" alt="ConceptMix Results">
55
+ </details>
56
+
57
+ <details>
58
+ <summary>GenEval 2 (<a href="[https://arxiv.org/abs/2512.16853](https://arxiv.org/abs/2512.16853)">Link</a>)</summary>
59
+ <br>
60
+ <img src="./GenEval2_results.png" alt="GenEval 2 Results">
61
+ </details>
62
+
63
+ <details>
64
+ <summary>T2I-CompBench (<a href="[https://arxiv.org/pdf/2307.06350v2](https://arxiv.org/pdf/2307.06350v2)">Link</a>)</summary>
65
+ <br>
66
+ <img src="./T2I-CompBench_results.png" alt="T2I-CompBench Results">
67
+ </details>
68
+
69
+ ## ๐Ÿ› ๏ธ Training Details
70
+ - **Base Model:** FLUX.1-dev
71
+ - **Algorithm:** Correlation-Weighted Multi-Reward Optimization (CMO)
72
+ - **Precision:** bfloat16
73
+
74
+ ## ๐Ÿ“œ Citation
75
+
76
+ If you find this model useful for your research, please cite:
77
+
78
+ ```bibtex
79
+ @article{wi2026correlation,
80
+ title={Correlation-Weighted Multi-Reward Optimization for Compositional Generation},
81
+ author={Wi, Jungmyung and Kim, Hyunsoo and Kim, Donghyun},
82
+ journal={arXiv preprint arXiv:2603.18528},
83
+ year={2026}
84
+ }
85
+ ```
T2I-CompBench_results.png ADDED
adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "FluxTransformer2DModel",
5
+ "parent_library": "diffusers.models.transformers.transformer_flux"
6
+ },
7
+ "base_model_name_or_path": null,
8
+ "bias": "none",
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": "gaussian",
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_dropout": 0.0,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 64,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "attn.add_v_proj",
27
+ "ff_context.net.0.proj",
28
+ "attn.add_k_proj",
29
+ "attn.to_q",
30
+ "ff.net.0.proj",
31
+ "ff_context.net.2",
32
+ "attn.to_out.0",
33
+ "attn.to_v",
34
+ "ff.net.2",
35
+ "attn.add_q_proj",
36
+ "attn.to_k",
37
+ "attn.to_add_out"
38
+ ],
39
+ "task_type": null,
40
+ "use_dora": false,
41
+ "use_rslora": false
42
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95f21e03b669d3c00356c487ad15018fb523e28c0d5b095a8269974b5b96f931
3
+ size 358709456
conceptmix_results.png ADDED

Git LFS Details

  • SHA256: 89f8f19b40762f1d044a1ccf06aac413d7d73fac9a7cb1fbf0e26e90bbb760b5
  • Pointer size: 131 Bytes
  • Size of remote file: 246 kB