NanoLLM Qwen v3.1
NanoLLM v3.1 artifacts are compact overlay artifacts for Qwen2.5 models. The loader starts from the base model in bitsandbytes 8-bit mode, then replaces the modules that passed the NanoLLM cascade with TrueQuantLinear modules.
Validated Artifacts
| Model | Artifact | Zip size | Gate | Avg cosine | Min cosine | Locked / 8-bit pending |
|---|---|---|---|---|---|---|
| Qwen2.5-3B-Instruct | final_artifact_3B.zip |
799,189,680 bytes | PASS | 0.990625 | 0.984375 | 143 / 109 |
| Qwen2.5-7B-Instruct | final_artifact_7B.zip |
891,419,698 bytes | PASS | 0.990625 | 0.98046875 | 66 / 130 |
| Qwen2.5-14B-Instruct | final_artifact_Qwen2.5-14B-Instruct_pruned_pass.zip |
1,482,019,132 bytes | PASS | 0.990625 | 0.98046875 | 76 / 260 |
The current release gate checks average next-token-logit cosine similarity against the 8-bit reference: avg >= 0.99. Minimum cosine is reported as a diagnostic.
Quick Start
from load_artifact import load_artifact
model, tokenizer, spec = load_artifact("final_artifact_Qwen2.5-14B-Instruct")
prompt = "Write a Python function to sort a list using bubble sort."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=160, do_sample=False)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Requirements:
pip install torch transformers accelerate bitsandbytes safetensors
Runtime Notes
build_reference_mode:8bitreference_scope:original_baselinepending_policy:leave_in_base_8bitNANO_LOAD_4BIT=1can be used experimentally to load the base model in 4-bit, but the release tests use 8-bit.
License
The NanoLLM quantization pipeline is proprietary/internal. Generated artifacts are published for research and evaluation subject to the repository license terms.
- Downloads last month
- 573