End of training
Browse files- README.md +2 -2
- config.json +9 -3
- diffusion_pytorch_model.bin +2 -2
- images_0.png +2 -2
- images_1.png +2 -2
README.md
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
|
| 2 |
---
|
| 3 |
license: creativeml-openrail-m
|
| 4 |
-
base_model:
|
| 5 |
tags:
|
| 6 |
- stable-diffusion
|
| 7 |
- stable-diffusion-diffusers
|
|
@@ -13,7 +13,7 @@ inference: true
|
|
| 13 |
|
| 14 |
# controlnet-razgaon/model-res
|
| 15 |
|
| 16 |
-
These are controlnet weights trained on
|
| 17 |
You can find some example images below.
|
| 18 |
prompt: a close up of a person wearing a black shirt
|
| 19 |

|
|
|
|
| 1 |
|
| 2 |
---
|
| 3 |
license: creativeml-openrail-m
|
| 4 |
+
base_model: stabilityai/stable-diffusion-2-1-base
|
| 5 |
tags:
|
| 6 |
- stable-diffusion
|
| 7 |
- stable-diffusion-diffusers
|
|
|
|
| 13 |
|
| 14 |
# controlnet-razgaon/model-res
|
| 15 |
|
| 16 |
+
These are controlnet weights trained on stabilityai/stable-diffusion-2-1-base with new type of conditioning.
|
| 17 |
You can find some example images below.
|
| 18 |
prompt: a close up of a person wearing a black shirt
|
| 19 |

|
config.json
CHANGED
|
@@ -1,8 +1,14 @@
|
|
| 1 |
{
|
| 2 |
"_class_name": "ControlNetModel",
|
| 3 |
"_diffusers_version": "0.17.0.dev0",
|
|
|
|
| 4 |
"act_fn": "silu",
|
| 5 |
-
"attention_head_dim":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
"block_out_channels": [
|
| 7 |
320,
|
| 8 |
640,
|
|
@@ -17,7 +23,7 @@
|
|
| 17 |
256
|
| 18 |
],
|
| 19 |
"controlnet_conditioning_channel_order": "rgb",
|
| 20 |
-
"cross_attention_dim":
|
| 21 |
"down_block_types": [
|
| 22 |
"CrossAttnDownBlock2D",
|
| 23 |
"CrossAttnDownBlock2D",
|
|
@@ -38,5 +44,5 @@
|
|
| 38 |
"projection_class_embeddings_input_dim": null,
|
| 39 |
"resnet_time_scale_shift": "default",
|
| 40 |
"upcast_attention": false,
|
| 41 |
-
"use_linear_projection":
|
| 42 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"_class_name": "ControlNetModel",
|
| 3 |
"_diffusers_version": "0.17.0.dev0",
|
| 4 |
+
"_name_or_path": "model-res/checkpoint-2000",
|
| 5 |
"act_fn": "silu",
|
| 6 |
+
"attention_head_dim": [
|
| 7 |
+
5,
|
| 8 |
+
10,
|
| 9 |
+
20,
|
| 10 |
+
20
|
| 11 |
+
],
|
| 12 |
"block_out_channels": [
|
| 13 |
320,
|
| 14 |
640,
|
|
|
|
| 23 |
256
|
| 24 |
],
|
| 25 |
"controlnet_conditioning_channel_order": "rgb",
|
| 26 |
+
"cross_attention_dim": 1024,
|
| 27 |
"down_block_types": [
|
| 28 |
"CrossAttnDownBlock2D",
|
| 29 |
"CrossAttnDownBlock2D",
|
|
|
|
| 44 |
"projection_class_embeddings_input_dim": null,
|
| 45 |
"resnet_time_scale_shift": "default",
|
| 46 |
"upcast_attention": false,
|
| 47 |
+
"use_linear_projection": true
|
| 48 |
}
|
diffusion_pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7badbb3b93231c4795fbe35ecb258a469de2b9a1eddf38c5fd9b4d295f81ad14
|
| 3 |
+
size 1457056057
|
images_0.png
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
images_1.png
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|