Update README with finetuning details and credits
Browse files
README.md
CHANGED
|
@@ -4,13 +4,14 @@ tags:
|
|
| 4 |
- diffusers
|
| 5 |
- ddpm
|
| 6 |
- unconditional-image-generation
|
|
|
|
| 7 |
library_name: diffusers
|
| 8 |
pipeline_tag: unconditional-image-generation
|
| 9 |
---
|
| 10 |
|
| 11 |
# ddpm-landscape
|
| 12 |
|
| 13 |
-
A DDPM
|
| 14 |
|
| 15 |
## Usage
|
| 16 |
|
|
@@ -29,3 +30,40 @@ image = ddpm().images[0]
|
|
| 29 |
# save image
|
| 30 |
image.save("ddpm_generated_image.png")
|
| 31 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
- diffusers
|
| 5 |
- ddpm
|
| 6 |
- unconditional-image-generation
|
| 7 |
+
- landscape
|
| 8 |
library_name: diffusers
|
| 9 |
pipeline_tag: unconditional-image-generation
|
| 10 |
---
|
| 11 |
|
| 12 |
# ddpm-landscape
|
| 13 |
|
| 14 |
+
A 256x256 unconditional DDPM that generates natural landscape images. Full fine-tune of [`google/ddpm-church-256`](https://huggingface.co/google/ddpm-church-256) on the **Landscapes HQ (LHQ)** dataset.
|
| 15 |
|
| 16 |
## Usage
|
| 17 |
|
|
|
|
| 30 |
# save image
|
| 31 |
image.save("ddpm_generated_image.png")
|
| 32 |
```
|
| 33 |
+
|
| 34 |
+
## Base model
|
| 35 |
+
|
| 36 |
+
- **`google/ddpm-church-256`** — original 256x256 DDPM by Ho et al. All credit for the base architecture and pretrained weights goes to the original authors.
|
| 37 |
+
|
| 38 |
+
## Dataset
|
| 39 |
+
|
| 40 |
+
- **Landscapes HQ (LHQ)**, 256x256 split, from *ALIS — Aligning Latent and Image Spaces to Connect the Unconnectable* (Skorokhodov et al., 2021).
|
| 41 |
+
- Project / data: https://github.com/universome/alis
|
| 42 |
+
|
| 43 |
+
```bibtex
|
| 44 |
+
@article{ALIS,
|
| 45 |
+
title = {Aligning Latent and Image Spaces to Connect the Unconnectable},
|
| 46 |
+
author = {Skorokhodov, Ivan and Sotnikov, Grigorii and Elhoseiny, Mohamed},
|
| 47 |
+
journal = {arXiv preprint arXiv:2104.06954},
|
| 48 |
+
year = {2021}
|
| 49 |
+
}
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
## Fine-tuning
|
| 53 |
+
|
| 54 |
+
| | |
|
| 55 |
+
|---|---|
|
| 56 |
+
| Base model | `google/ddpm-church-256` |
|
| 57 |
+
| Dataset | LHQ (256x256) |
|
| 58 |
+
| Epochs | 50 |
|
| 59 |
+
| Batch size | 32 |
|
| 60 |
+
| Optimizer | AdamW |
|
| 61 |
+
| Learning rate | 1e-5 (cosine schedule, 500 warmup steps) |
|
| 62 |
+
| Loss | MSE on predicted noise |
|
| 63 |
+
| Augmentation | Random horizontal flip |
|
| 64 |
+
|
| 65 |
+
## Acknowledgements
|
| 66 |
+
|
| 67 |
+
- Base weights: Google / the original DDPM authors (Ho, Jain, Abbeel, 2020).
|
| 68 |
+
- Dataset: Skorokhodov et al., authors of ALIS / LHQ — https://github.com/universome/alis
|
| 69 |
+
- Built with [`diffusers`](https://github.com/huggingface/diffusers).
|