Qwen3-4B_LoopUS / README.md
Thrillcrazyer's picture
Improve model card: add pipeline tag, paper link, and code link (#1)
8e1b3a7
metadata
base_model:
  - Qwen/Qwen3-4B-Base
datasets:
  - HuggingFaceFW/fineweb-edu
license: apache-2.0
model_name: Qwen3_1.7B_LoopUS_SFT
pipeline_tag: text-generation
tags:
  - LoopUS
  - LoopedTrasnformers

LoopUS:
Recasting Pretrained LLMs into Looped Latent Refinement Models

BAELAB, Pusan National University, Busan, Korea
DOLAB, Changwon National University, Changwon, Korea

Taekhyun Park1, Yongjae Lee1, Dohee Kim2, Hyerim Bae1,โ€ 

๐ŸŒŸ Github | ๐ŸŒ Project Page | ๐Ÿ“„ Paper

Abstract

Looped computation shows promise in improving the reasoning-oriented performance of LLMs by scaling test-time compute. Looped Depth Up-Scaling (LoopUS) is a post-training framework that converts a standard pretrained LLM into a looped architecture. LoopUS recasts the pretrained LLM into an encoder, a looped reasoning block, and a decoder. It improves reasoning-oriented performance without extending the generated traces or requiring recurrent training from scratch.

QuickStart

To use this model, clone the official repository and run the chat interface:

git clone https://github.com/Thrillcrazyer/LoopUS.git
cd LoopUS
uv sync
uv run chat.py --model-name Thrillcrazyer/Qwen3_1.7B_LoopUS_SFT

Illustration of LoopUS

Citation

If you find LoopUS useful in your research, please cite:

@article{park2026loopus,
  title={LoopUS: Recasting Pretrained LLMs into Looped Latent Refinement Models},
  author={Park, Taekhyun and Lee, Yongjae and Kim, Dohee and Bae, Hyerim},
  journal={arXiv preprint arXiv:2605.11011},
  year={2026}
}