File size: 5,295 Bytes
2611973 7f7272e 2611973 7f7272e f6dcedd 7f7272e 4052a4c 27ede4b 44853a0 7f7272e 6348fde 7f7272e ff43e43 7f7272e 44853a0 7f7272e 76afa55 7f7272e aad58c8 7f7272e 4052a4c 44853a0 7f7272e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 | ---
license: mit
language:
- en
tags:
- image-restoration
- super-resolution
- visual-autoregressive
- pytorch
---
# VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution (ICLR 2026)
<div align="center">
<img src="./assets/logo-new-2.jpg" alt="VARestorer Logo" width="380"/>
<h3>
📄 <a href="https://openreview.net/pdf?id=T2Oihh7zN8">Paper</a>
📝 <a href="http://arxiv.org/abs/2604.21450">arXiv</a>
🏠 <a href="https://eternalevan.github.io/VARestorer-proj/">Project Page</a>
💻 <a href="https://github.com/EternalEvan/VARestorer">Code</a>
</h3>
</div>
<div align="center">
**[Yixuan Zhu*](https://eternalevan.github.io/), [Shilin Ma*](https://github.com/cyp336/), [Haolin Wang](https://howlin-wang.github.io/), [Ao Li](https://rammusleo.github.io/), Yanzhe Jing, [Yansong Tang†](https://andytang15.github.io/), [Lei Chen](https://scholar.google.com/citations?user=8bMh-FQAAAAJ&hl=zh-CN&oi=sra), [Jiwen Lu](http://ivg.au.tsinghua.edu.cn/Jiwen_Lu/), [Jie Zhou](https://scholar.google.com/citations?user=6a79aPwAAAAJ&hl=en)**
(* Equal contribution † Corresponding author)
Tsinghua University
</div>
VARestorer is the official Hugging Face model repository for the ICLR 2026 paper **"VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution."** It distills a pre-trained text-to-image visual autoregressive (VAR) model into a **single-step** real-world image super-resolution system.
This Hugging Face repository includes the released checkpoint assets together with a runnable snapshot of the official codebase for convenient download and inference. The primary development home remains the official [GitHub repository](https://github.com/EternalEvan/VARestorer).
## Real-World Restoration at a Glance
|  |  |  |
|:---:|:---:|:---:|
| Street Scene | Landscape | Corgi Portrait |
<p align="center">
<sub>
Left half: <b>real degraded input</b> | Right half: <b>VARestorer</b> one-step output. <br/>
Want to drag the divider yourself? → <a href="https://eternalevan.github.io/VARestorer-proj/">Try the interactive slider on the project page</a>.
</sub>
</p>
| **1 step** | **0.23 s** | **~10× faster** | **27.3 M params** |
| :---: | :---: | :---: | :---: |
| one-pass inference | per 512×512 image | than VAR baseline | trainable (1.2% of total) |
## Pipeline

## Download and Use
You can either clone this Hugging Face repository directly or use the primary GitHub repository. The commands below assume the usual GitHub workflow, but the same directory layout is now mirrored here as well.
1. Clone the repository and install the dependencies:
```bash
# Option A: clone the primary GitHub repository
git clone https://github.com/EternalEvan/VARestorer.git
cd VARestorer
pip install -r requirements.txt
pip install --no-build-isolation git+https://github.com/cloneofsimo/lora.git
pip install --no-build-isolation flash_attn==2.8.3
```
If you prefer to clone the Hugging Face mirror instead, use:
```bash
git clone https://huggingface.co/EternalEvan/VARestorer
cd VARestorer
pip install -r requirements.txt
pip install --no-build-isolation git+https://github.com/cloneofsimo/lora.git
pip install --no-build-isolation flash_attn==2.8.3
```
2. Download the main checkpoint from this repository ([weights](https://huggingface.co/EvanEternal/VARestorer/tree/main/weights)) or from [Google Drive](https://drive.google.com/file/d/1NkwlvNfr7nOkN45VWmO-PXbJZ8Nkt2_l/view?usp=drive_link):
```bash
huggingface-cli download EvanEternal/VARestorer varestorer.pth --local-dir ./weights
```
3. Download the additional dependencies required by the official release:
- [`google/flan-t5-xl`](https://huggingface.co/google/flan-t5-xl) into `./weights/flan-t5-xl`
- [`lxq007/DiffBIR`](https://huggingface.co/lxq007/DiffBIR/blob/main/general_swinir_v1.ckpt) as `./weights/general_swinir_v1.ckpt`
- [`FoundationVision/Infinity`](https://huggingface.co/FoundationVision/Infinity/blob/main/infinity_vae_d32reg.pth) as `./weights/infinity_vae_d32reg.pth`
4. Run inference:
```bash
bash scripts/infer.sh
```
For the latest updates, issue tracking, and future development, please refer to the [official GitHub repository](https://github.com/EternalEvan/VARestorer).
## Links
- [Paper (OpenReview)](https://openreview.net/forum?id=T2Oihh7zN8)
- [arXiv](http://arxiv.org/abs/2604.21450)
- [Project Page](https://eternalevan.github.io/VARestorer-proj/)
- [Code Repository](https://github.com/EternalEvan/VARestorer)
## Citation
```bibtex
@inproceedings{zhu2026varestorer,
title = {VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution},
author = {Zhu, Yixuan and Ma, Shilin and Wang, Haolin and Li, Ao and
Jing, Yanzhe and Tang, Yansong and Chen, Lei and Lu, Jiwen and Zhou, Jie},
booktitle = {International Conference on Learning Representations (ICLR)},
year = {2026},
url = {https://openreview.net/forum?id=T2Oihh7zN8}
}
```
|