{ "architectures": [ "MagicBERT" ], "attention_dropout": 0.15, "auto_map": { "AutoConfig": "config.MagicBERTConfig", "AutoModel": "modeling.MagicBERT" }, "d_model": 128, "dim_feed_forward": 341, "dtype": "float32", "embedding_dropout": 0.15, "mask_token_id": 0, "model_type": "magicBERT", "num_attention_heads": 8, "num_encoder_layers": 4, "pad_token_id": 1, "seq_len": 100, "tie_embeddings": true, "transformers_version": "4.57.3", "vocab_size": 36476 }