{ "num_conversations": 3267, "model_config": { "vocab_size": 4000, "d_model": 256, "n_heads": 8, "n_layers": 6, "d_ff": 1024, "dropout": 0.1, "max_len": 512 }, "tokenizer_vocab_size": 4000 }