File size: 5,295 Bytes
2611973
 
7f7272e
 
 
 
 
 
 
2611973
7f7272e
 
 
 
f6dcedd
7f7272e
 
 
4052a4c
27ede4b
44853a0
7f7272e
 
 
 
 
 
 
6348fde
7f7272e
 
 
 
 
 
 
 
 
 
 
 
ff43e43
 
 
7f7272e
 
 
 
44853a0
7f7272e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76afa55
7f7272e
 
aad58c8
7f7272e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4052a4c
44853a0
7f7272e
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
---

license: mit
language:
- en
tags:
- image-restoration
- super-resolution
- visual-autoregressive
- pytorch
---


# VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution (ICLR 2026)

<div align="center">
  <img src="./assets/logo-new-2.jpg" alt="VARestorer Logo" width="380"/>
  <h3>
    📄 <a href="https://openreview.net/pdf?id=T2Oihh7zN8">Paper</a>

    &nbsp;&nbsp;

    📝 <a href="http://arxiv.org/abs/2604.21450">arXiv</a>

    &nbsp;&nbsp;

    🏠 <a href="https://eternalevan.github.io/VARestorer-proj/">Project Page</a>

    &nbsp;&nbsp;

    💻 <a href="https://github.com/EternalEvan/VARestorer">Code</a>

  </h3>

</div>


<div align="center">

**[Yixuan Zhu*](https://eternalevan.github.io/), [Shilin Ma*](https://github.com/cyp336/), [Haolin Wang](https://howlin-wang.github.io/), [Ao Li](https://rammusleo.github.io/), Yanzhe Jing, [Yansong Tang†](https://andytang15.github.io/), [Lei Chen](https://scholar.google.com/citations?user=8bMh-FQAAAAJ&hl=zh-CN&oi=sra), [Jiwen Lu](http://ivg.au.tsinghua.edu.cn/Jiwen_Lu/), [Jie Zhou](https://scholar.google.com/citations?user=6a79aPwAAAAJ&hl=en)**

(* Equal contribution &nbsp; † Corresponding author)

Tsinghua University
</div>

VARestorer is the official Hugging Face model repository for the ICLR 2026 paper **"VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution."** It distills a pre-trained text-to-image visual autoregressive (VAR) model into a **single-step** real-world image super-resolution system.

This Hugging Face repository includes the released checkpoint assets together with a runnable snapshot of the official codebase for convenient download and inference. The primary development home remains the official [GitHub repository](https://github.com/EternalEvan/VARestorer).

## Real-World Restoration at a Glance

| ![](./assets/teaser_car.webp) | ![](./assets/teaser_field.webp) | ![](./assets/teaser_corgi.webp) |
|:---:|:---:|:---:|
| Street Scene | Landscape | Corgi Portrait |

<p align="center">
  <sub>
    Left half: <b>real degraded input</b> &nbsp;|&nbsp; Right half: <b>VARestorer</b> one-step output. <br/>

    Want to drag the divider yourself? &rarr; <a href="https://eternalevan.github.io/VARestorer-proj/">Try the interactive slider on the project page</a>.

  </sub>

</p>


| **1 step** | **0.23 s** | **~10× faster** | **27.3 M params** |
| :---: | :---: | :---: | :---: |
| one-pass inference | per 512×512 image | than VAR baseline | trainable (1.2% of total) |

## Pipeline

![](./assets/pipeline.png)

## Download and Use

You can either clone this Hugging Face repository directly or use the primary GitHub repository. The commands below assume the usual GitHub workflow, but the same directory layout is now mirrored here as well.

1. Clone the repository and install the dependencies:

```bash

# Option A: clone the primary GitHub repository

git clone https://github.com/EternalEvan/VARestorer.git

cd VARestorer

pip install -r requirements.txt

pip install --no-build-isolation git+https://github.com/cloneofsimo/lora.git

pip install --no-build-isolation flash_attn==2.8.3

```

If you prefer to clone the Hugging Face mirror instead, use:

```bash

git clone https://huggingface.co/EternalEvan/VARestorer

cd VARestorer

pip install -r requirements.txt

pip install --no-build-isolation git+https://github.com/cloneofsimo/lora.git

pip install --no-build-isolation flash_attn==2.8.3

```

2. Download the main checkpoint from this repository ([weights](https://huggingface.co/EvanEternal/VARestorer/tree/main/weights)) or from [Google Drive](https://drive.google.com/file/d/1NkwlvNfr7nOkN45VWmO-PXbJZ8Nkt2_l/view?usp=drive_link):

```bash

huggingface-cli download EvanEternal/VARestorer varestorer.pth --local-dir ./weights

```

3. Download the additional dependencies required by the official release:

- [`google/flan-t5-xl`](https://huggingface.co/google/flan-t5-xl) into `./weights/flan-t5-xl`
- [`lxq007/DiffBIR`](https://huggingface.co/lxq007/DiffBIR/blob/main/general_swinir_v1.ckpt) as `./weights/general_swinir_v1.ckpt`
- [`FoundationVision/Infinity`](https://huggingface.co/FoundationVision/Infinity/blob/main/infinity_vae_d32reg.pth) as `./weights/infinity_vae_d32reg.pth`

4. Run inference:

```bash

bash scripts/infer.sh

```

For the latest updates, issue tracking, and future development, please refer to the [official GitHub repository](https://github.com/EternalEvan/VARestorer).

## Links

- [Paper (OpenReview)](https://openreview.net/forum?id=T2Oihh7zN8)
- [arXiv](http://arxiv.org/abs/2604.21450)
- [Project Page](https://eternalevan.github.io/VARestorer-proj/)
- [Code Repository](https://github.com/EternalEvan/VARestorer)

## Citation

```bibtex

@inproceedings{zhu2026varestorer,

  title     = {VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution},

  author    = {Zhu, Yixuan and Ma, Shilin and Wang, Haolin and Li, Ao and

               Jing, Yanzhe and Tang, Yansong and Chen, Lei and Lu, Jiwen and Zhou, Jie},

  booktitle = {International Conference on Learning Representations (ICLR)},

  year      = {2026},

  url       = {https://openreview.net/forum?id=T2Oihh7zN8}

}

```