File size: 3,165 Bytes
00d986f
c8e514a
 
00d986f
 
c8e514a
00d986f
c8e514a
59f1229
 
 
00d986f
c8e514a
00d986f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c8e514a
 
cdbab9d
00d986f
 
c8e514a
 
 
 
 
 
 
00d986f
c8e514a
00d986f
c8e514a
00d986f
c8e514a
 
 
 
 
 
 
00d986f
 
 
 
c8e514a
00d986f
 
c8e514a
 
 
 
00d986f
c8e514a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
---
base_model:
- microsoft/phi-4
datasets:
- HuggingFaceFW/fineweb-edu
license: apache-2.0
model_name: Qwen3_1.7B_LoopUS_SFT
pipeline_tag: text-generation
tags:
- LoopUS
- LoopedTransformers
---

<div align="center">
<h1>LoopUS: <br> Recasting Pretrained LLMs into Looped Latent Refinement Models</h1>
</div>

<p align="center">
  <a href="https://pnubaelab.github.io/"><b>BAELAB</b></a>, Pusan National University, Busan, Korea <br>
  <a href="https://aidoheekim.github.io/"><b>DOLAB</b></a>, Changwon National University, Changwon, Korea
</p>

<p align="center">
  <a href="https://thrillcrazyer.github.io/" target="_blank"><strong>Taekhyun Park</strong></a><sup>1</sup>,
  <a href="https://yongzzai.com/" target="_blank"><strong>Yongjae Lee</strong></a><sup>1</sup>,
  <a href="https://aidoheekim.github.io/" target="_blank"><strong>Dohee Kim</strong></a><sup>2</sup>,
  <a href="https://pnubaelab.github.io/" target="_blank"><strong>Hyerim Bae</string></a><sup>1,&dagger;</sup>
</p>

<p align="center">
  <a href="https://github.com/Thrillcrazyer/LoopUS"><b>๐ŸŒŸ Github</b></a> |
  <a href="https://thrillcrazyer.github.io/LoopUS"><b>๐ŸŒ Project Page</b></a> |
  <a href="https://arxiv.org/abs/2605.11011"><b>๐Ÿ“„ Paper</b></a> 
</p>

# Overview

**Looped Depth Up-Scaling** (LoopUS) is a post-training framework that converts a standard pretrained LLM into a looped architecture. LoopUS recasts the pretrained LLM into an encoder, a looped reasoning block, and a decoder. It operationalizes this latent-refinement architecture through:
1. **Block Decomposition:** Recasts a pretrained transformer into a reusable latent-refinement architecture.
2. **Input-Dependent Selective Gate:** Adaptively controls hidden state propagation to mitigate drift.
3. **Random Deep Supervision:** Enables memory-efficient learning over long recursive horizons.
4. **Confidence Head:** Allows for adaptive early exiting during inference.

Through stable latent looping, LoopUS improves reasoning-oriented performance without extending the generated traces or requiring recurrent training from scratch.

# Illustration of LoopUS

<div align="center">
<img src="https://raw.githubusercontent.com/Thrillcrazyer/LoopUS/main/assets/Framework.png" width="800"/>
</div>

# Quick Start

To use this model, please follow the installation instructions in the [official repository](https://github.com/Thrillcrazyer/LoopUS):

```bash
git clone https://github.com/Thrillcrazyer/LoopUS.git
cd LoopUS
uv sync
```

### Chatting Mode
```bash
uv run chat.py --model-name Thrillcrazyer/Qwen3_1.7B_LoopUS_SFT
```

### Qualitative Generation
```bash
uv run LoopUS-generate \
    --model-name microsoft/phi-4 \
    --decomposed-model Thrillcrazyer/Qwen3_1.7B_LoopUS_SFT \
    --prompt "The meaning of life is" \
    --n-recursion 8
```

# Citation

```bibtex
@misc{park2024loopus,
      title={LoopUS: Recasting Pretrained LLMs into Looped Latent Refinement Models}, 
      author={Taekhyun Park and Yongjae Lee and Dohee Kim and Hyerim Bae},
      year={2024},
      eprint={2605.11011},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2605.11011}, 
}
```