Kronos-base (MLX)
Apple MLX port of NeoQuasar/Kronos-base โ the 102M-parameter Kronos variant. d_model=832, n_layers=12, max_context=512. Pair with gxcsoccer/kronos-mlx-tokenizer-base.
Usage
from kronos_mlx import Kronos, KronosTokenizer, KronosPredictor
tokenizer = KronosTokenizer.from_pretrained("gxcsoccer/kronos-mlx-tokenizer-base")
model = Kronos.from_pretrained("gxcsoccer/kronos-mlx-base")
predictor = KronosPredictor(model, tokenizer, max_context=512)
For 8-bit Linear weight quantization (390 MB โ ~115 MB, -71 %):
model = Kronos.from_pretrained("gxcsoccer/kronos-mlx-base", bits=8)
8-bit on Kronos-base is much higher fidelity than on Kronos-small thanks to the larger model's redundancy โ recommended for memory-constrained Apple Silicon.
Original
- Upstream: shiyu-coder/Kronos (AAAI 2026)
- PyTorch weights: NeoQuasar/Kronos-base
- MLX port: github.com/gxcsoccer/kronos-mlx
- Downloads last month
- -
Hardware compatibility
Log In to add your hardware
Quantized
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for gxcsoccer/kronos-mlx-base
Base model
NeoQuasar/Kronos-base