A newer version of this model is available: qikp/kite-2.6-13m
Kite
🎉 You are looking at Kite 1.6, which is now trained using pika!
Kite is a small, trained, 1 million parameter language model, without any special optimizations.
Training
It was trained on this dataset using 20000 steps, 1 epoch, 1 batch size, and the pika tokenizer.
Limitations
Due to its size, the model is not suitable for production workloads.
- Downloads last month
- 4
Model tree for qikp/kite-1.6-1m
Base model
hmellor/tiny-random-LlamaForCausalLM