Ezekiel999/AksaraLLM-20B-Instruct (smoke-test checkpoint)
This is NOT the production 20B model. It is a randomly-initialized
tiny preset (2 layers, 64-dim, vocab 256) pushed from a Devin
scaffolding session to validate the aksaraLLMModel.save_pretrained โ
HF โ aksaraLLMModel.from_pretrained round-trip.
The real 20B model (42 layers, 6144-dim, vocab 131 072) must be trained
from random initialisation on a TPU v5p pod using
aksara-train/scripts/train_20b_pretrain.py.
- Downloads last month
- 12
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support