metadata
license: apache-2.0
language: id
library_name: transformers
tags:
- aksarallm
- indonesian
- from-scratch
- smoke-test
Ezekiel999/AksaraLLM-20B-Instruct (smoke-test checkpoint)
This is NOT the production 20B model. It is a randomly-initialized
tiny preset (2 layers, 64-dim, vocab 256) pushed from a Devin
scaffolding session to validate the aksaraLLMModel.save_pretrained →
HF → aksaraLLMModel.from_pretrained round-trip.
The real 20B model (42 layers, 6144-dim, vocab 131 072) must be trained
from random initialisation on a TPU v5p pod using
aksara-train/scripts/train_20b_pretrain.py.