File size: 3,165 Bytes
00d986f c8e514a 00d986f c8e514a 00d986f c8e514a 59f1229 00d986f c8e514a 00d986f c8e514a cdbab9d 00d986f c8e514a 00d986f c8e514a 00d986f c8e514a 00d986f c8e514a 00d986f c8e514a 00d986f c8e514a 00d986f c8e514a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 | ---
base_model:
- microsoft/phi-4
datasets:
- HuggingFaceFW/fineweb-edu
license: apache-2.0
model_name: Qwen3_1.7B_LoopUS_SFT
pipeline_tag: text-generation
tags:
- LoopUS
- LoopedTransformers
---
<div align="center">
<h1>LoopUS: <br> Recasting Pretrained LLMs into Looped Latent Refinement Models</h1>
</div>
<p align="center">
<a href="https://pnubaelab.github.io/"><b>BAELAB</b></a>, Pusan National University, Busan, Korea <br>
<a href="https://aidoheekim.github.io/"><b>DOLAB</b></a>, Changwon National University, Changwon, Korea
</p>
<p align="center">
<a href="https://thrillcrazyer.github.io/" target="_blank"><strong>Taekhyun Park</strong></a><sup>1</sup>,
<a href="https://yongzzai.com/" target="_blank"><strong>Yongjae Lee</strong></a><sup>1</sup>,
<a href="https://aidoheekim.github.io/" target="_blank"><strong>Dohee Kim</strong></a><sup>2</sup>,
<a href="https://pnubaelab.github.io/" target="_blank"><strong>Hyerim Bae</string></a><sup>1,†</sup>
</p>
<p align="center">
<a href="https://github.com/Thrillcrazyer/LoopUS"><b>๐ Github</b></a> |
<a href="https://thrillcrazyer.github.io/LoopUS"><b>๐ Project Page</b></a> |
<a href="https://arxiv.org/abs/2605.11011"><b>๐ Paper</b></a>
</p>
# Overview
**Looped Depth Up-Scaling** (LoopUS) is a post-training framework that converts a standard pretrained LLM into a looped architecture. LoopUS recasts the pretrained LLM into an encoder, a looped reasoning block, and a decoder. It operationalizes this latent-refinement architecture through:
1. **Block Decomposition:** Recasts a pretrained transformer into a reusable latent-refinement architecture.
2. **Input-Dependent Selective Gate:** Adaptively controls hidden state propagation to mitigate drift.
3. **Random Deep Supervision:** Enables memory-efficient learning over long recursive horizons.
4. **Confidence Head:** Allows for adaptive early exiting during inference.
Through stable latent looping, LoopUS improves reasoning-oriented performance without extending the generated traces or requiring recurrent training from scratch.
# Illustration of LoopUS
<div align="center">
<img src="https://raw.githubusercontent.com/Thrillcrazyer/LoopUS/main/assets/Framework.png" width="800"/>
</div>
# Quick Start
To use this model, please follow the installation instructions in the [official repository](https://github.com/Thrillcrazyer/LoopUS):
```bash
git clone https://github.com/Thrillcrazyer/LoopUS.git
cd LoopUS
uv sync
```
### Chatting Mode
```bash
uv run chat.py --model-name Thrillcrazyer/Qwen3_1.7B_LoopUS_SFT
```
### Qualitative Generation
```bash
uv run LoopUS-generate \
--model-name microsoft/phi-4 \
--decomposed-model Thrillcrazyer/Qwen3_1.7B_LoopUS_SFT \
--prompt "The meaning of life is" \
--n-recursion 8
```
# Citation
```bibtex
@misc{park2024loopus,
title={LoopUS: Recasting Pretrained LLMs into Looped Latent Refinement Models},
author={Taekhyun Park and Yongjae Lee and Dohee Kim and Hyerim Bae},
year={2024},
eprint={2605.11011},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2605.11011},
}
``` |