minillama-base / config.json
nishtahir's picture
Upload 9 files
d451d06 verified
raw
history blame contribute delete
432 Bytes
{
"architectures": [
"MiniLlama"
],
"auto_map": {
"AutoConfig": "config_minillama.MiniLlamaConfig",
"AutoModel": "modeling_minillama.MiniLlama"
},
"dim": 768,
"dropout": 0.1,
"dtype": "float32",
"max_seq_len": 1024,
"model_type": "mini-llama",
"multiple_of": 256,
"n_heads": 12,
"n_kv_heads": 12,
"n_layers": 24,
"norm_eps": 1e-05,
"transformers_version": "4.57.3",
"vocab_size": 50257
}