{ "architectures": [ "TokenSeparator" ], "bos_token_id": 1024, "codebook_decoder_layers": 4, "codebook_size": 1024, "d_model": 768, "dim_feedforward": 3072, "dropout": 0.1, "dtype": "float32", "latent_dim": 128, "model_type": "token_separator", "n_stems": 4, "nhead": 8, "num_layers": 8, "predict_n_q": 16, "rope_base": 10000.0, "tie_codebook_embeddings": false, "transformers_version": "5.2.0", "use_cache": false, "vocab_size": 1025 }