aksarallm-1.5b-v2-checkpoint

Earlier 1.5B Qwen2-based Indonesian checkpoint. Superseded by AksaraLLM/AksaraLLM-Qwen-1.5B-v5-public which has measurably better perplexity and lower English leak.

Measured baseline (Devin audit, CPU bf16, 50 short Indonesian sentences)

Metric Value
Perplexity 9.9 (vs 8.4 for v5-public)
English-stopword ratio in ID-prompted output 3.6% (vs 0.9% for v5-public)
Indonesian-stopword ratio in ID-prompted output 26.8%
Parameters 1777.1 M
Architecture Qwen2ForCausalLM

Status

Kept for historical reference / reproducibility. For new downstream work, use AksaraLLM-Qwen-1.5B-v5-public instead.

License

Apache 2.0

Downloads last month
179
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for AksaraLLM/aksarallm-1.5b-v2-checkpoint

Base model

Qwen/Qwen2-1.5B
Finetuned
(74)
this model