I-DLM-8B-LoRA-r128

LoRA adapter (rank=128) for I-DLM-8B, enabling lossless Relaxed Introspective Strided Decoding (R-ISD).

This model was presented in the paper Introspective Diffusion Language Models.

[Project Page] [Paper] [Code]

Overview

This LoRA adapter is used in the gated LoRA setup described in the I-DLM paper:

  • LoRA is active only on MASK positions during ISD decoding
  • Clean/verify positions use base model weights only
  • This produces output bit-for-bit identical to the base AR model (Qwen3-8B)

Usage

Note: This LoRA adapter is used with the I-DLM-8B base model via our SGLang-based serving pipeline. Direct loading via transformers is not currently supported for reproducing paper results. Please use the SGLang-based ISD pipeline for inference.

`

Downloads last month
81
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for yifanyu/I-DLM-8B-lora-r128

Finetuned
Qwen/Qwen3-8B
Finetuned
yifanyu/I-DLM-8B
Adapter
(1)
this model

Collection including yifanyu/I-DLM-8B-lora-r128

Paper for yifanyu/I-DLM-8B-lora-r128