nielsr HF Staff commited on
Commit
60e5c43
Β·
verified Β·
1 Parent(s): b362e5b

Improve model card metadata and content

Browse files

This PR updates the model card with the correct `pipeline_tag` to improve discoverability. It also adds a link to the project page and completes the sample usage snippet using the `diffusers` library.

Files changed (1) hide show
  1. README.md +42 -31
README.md CHANGED
@@ -1,24 +1,24 @@
1
  ---
2
- license: mit
3
  library_name: diffusers
 
 
4
  tags:
5
- - computed-tomography
6
- - ct-reconstruction
7
- - diffusion-model
8
- - latent-diffusion
9
- - inverse-problems
10
- - dm4ct
11
- - sparse-view-ct
12
  ---
13
 
14
  # Latent Diffusion Model – LoDoChallenge (DM4CT)
15
 
16
- This repository contains the pretrained **latent-space diffusion model** used in the
17
- **DM4CT: Benchmarking Diffusion Models for CT Reconstruction (ICLR 2026)** benchmark.
18
 
19
- πŸ”— Paper: https://openreview.net/forum?id=YE5scJekg5
20
- πŸ”— Arxiv: https://arxiv.org/abs/2602.18589
21
- πŸ”— Codebase: https://github.com/DM4CT/DM4CT
22
 
23
  ---
24
 
@@ -32,23 +32,19 @@ Unlike the pixel diffusion model, diffusion is performed in the latent space of
32
  - VQ-VAE (image encoder/decoder)
33
  - 2D UNet operating in latent space
34
  - **Input resolution (image space)**: 512 Γ— 512
35
- - **Latent resolution**: (insert latent size, e.g., 64 Γ— 64)
36
  - **Channels**: 1 (grayscale CT slice)
37
  - **Training objective**: Ξ΅-prediction (standard DDPM formulation)
38
  - **Noise schedule**: Linear beta schedule
39
  - **Training dataset**: Low Dose Grand Challenge (LoDoChallenge)
40
  - **Intensity normalization**: Rescaled to (-1, 1)
41
 
42
- The diffusion model operates purely in latent space and relies on the autoencoder for encoding and decoding.
43
-
44
- This model is intended to be combined with data-consistency correction for CT reconstruction.
45
 
46
  ---
47
 
48
  ## πŸ“Š Dataset: Low Dose Grand Challenge
49
 
50
- Source:
51
- https://www.aapm.org/grandchallenge/lowdosect/
52
 
53
  Preprocessing steps:
54
  - Train/test split
@@ -61,25 +57,40 @@ The model learns an unconditional latent prior over CT slices.
61
 
62
  ## 🧠 Training Details
63
 
64
- - Optimizer: AdamW
65
- - Learning rate: 1e-4
66
- - Batch size: (insert your batch size)
67
- - Training steps: (insert number of steps)
68
- - Hardware: NVIDIA A100 GPU
69
-
70
- Training scripts:
71
- - Latent diffusion: https://github.com/DM4CT/DM4CT/blob/main/train_latent.py
72
- - Autoencoder training: (insert if separate)
73
 
74
  ---
75
 
76
  ## πŸš€ Usage
77
 
78
  ```python
79
- from diffusers import LDMPipeline
 
80
 
81
- LDMPipeline = DiffusionPipeline.from_pretrained(
82
  "jiayangshi/lodochallenge_latent_diffusion"
83
  )
84
 
85
- pipeline.to("cuda")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  library_name: diffusers
3
+ license: mit
4
+ pipeline_tag: image-to-image
5
  tags:
6
+ - computed-tomography
7
+ - ct-reconstruction
8
+ - diffusion-model
9
+ - latent-diffusion
10
+ - inverse-problems
11
+ - dm4ct
12
+ - sparse-view-ct
13
  ---
14
 
15
  # Latent Diffusion Model – LoDoChallenge (DM4CT)
16
 
17
+ This repository contains the pretrained **latent-space diffusion model** used in the benchmark **DM4CT: Benchmarking Diffusion Models for Computed Tomography Reconstruction (ICLR 2026)**.
 
18
 
19
+ - **Paper:** [DM4CT: Benchmarking Diffusion Models for Computed Tomography Reconstruction](https://huggingface.co/papers/2602.18589)
20
+ - **Project Page:** [https://dm4ct.github.io/DM4CT/](https://dm4ct.github.io/DM4CT/)
21
+ - **Codebase:** [https://github.com/DM4CT/DM4CT](https://github.com/DM4CT/DM4CT)
22
 
23
  ---
24
 
 
32
  - VQ-VAE (image encoder/decoder)
33
  - 2D UNet operating in latent space
34
  - **Input resolution (image space)**: 512 Γ— 512
 
35
  - **Channels**: 1 (grayscale CT slice)
36
  - **Training objective**: Ξ΅-prediction (standard DDPM formulation)
37
  - **Noise schedule**: Linear beta schedule
38
  - **Training dataset**: Low Dose Grand Challenge (LoDoChallenge)
39
  - **Intensity normalization**: Rescaled to (-1, 1)
40
 
41
+ The diffusion model operates purely in latent space and relies on the autoencoder for encoding and decoding. This model is intended to be combined with data-consistency correction for CT reconstruction.
 
 
42
 
43
  ---
44
 
45
  ## πŸ“Š Dataset: Low Dose Grand Challenge
46
 
47
+ Source: [https://www.aapm.org/grandchallenge/lowdosect/](https://www.aapm.org/grandchallenge/lowdosect/)
 
48
 
49
  Preprocessing steps:
50
  - Train/test split
 
57
 
58
  ## 🧠 Training Details
59
 
60
+ - **Optimizer**: AdamW
61
+ - **Learning rate**: 1e-4
62
+ - **Hardware**: NVIDIA A100 GPU
63
+ - **Training scripts**: [train_latent.py](https://github.com/DM4CT/DM4CT/blob/main/train_latent.py)
 
 
 
 
 
64
 
65
  ---
66
 
67
  ## πŸš€ Usage
68
 
69
  ```python
70
+ from diffusers import DiffusionPipeline
71
+ import torch
72
 
73
+ pipeline = DiffusionPipeline.from_pretrained(
74
  "jiayangshi/lodochallenge_latent_diffusion"
75
  )
76
 
77
+ pipeline.to("cuda")
78
+
79
+ # Generate an unconditional CT slice prior
80
+ image = pipeline(batch_size=1).images[0]
81
+ image.save("reconstructed_slice.png")
82
+ ```
83
+
84
+ ---
85
+
86
+ ## Citation
87
+
88
+ ```bibtex
89
+ @inproceedings{shi2026dmct,
90
+ title={{DM}4{CT}: Benchmarking Diffusion Models for Computed Tomography Reconstruction},
91
+ author={Shi, Jiayang and Pelt, Dani{\"e}l M and Batenburg, K Joost},
92
+ booktitle={The Fourteenth International Conference on Learning Representations},
93
+ year={2026},
94
+ url={https://openreview.net/forum?id=YE5scJekg5}
95
+ }
96
+ ```