{ "architectures": [ "OpenAIPrivacyFilterForTokenClassification" ], "attention_bias": true, "attention_dropout": 0.0, "bos_token_id": null, "classifier_dropout": 0.0, "default_n_ctx": 128000, "dtype": "float16", "eos_token_id": 199999, "head_dim": 64, "hidden_act": "silu", "hidden_size": 640, "id2label": { "0": "O", "1": "B-COMPETITOR_MENTION", "2": "I-COMPETITOR_MENTION", "3": "E-COMPETITOR_MENTION", "4": "S-COMPETITOR_MENTION", "5": "B-PRIVACY_PII", "6": "I-PRIVACY_PII", "7": "E-PRIVACY_PII", "8": "S-PRIVACY_PII", "9": "B-PROFANITY", "10": "I-PROFANITY", "11": "E-PROFANITY", "12": "S-PROFANITY", "13": "B-FINANCIAL_ADVICE", "14": "I-FINANCIAL_ADVICE", "15": "E-FINANCIAL_ADVICE", "16": "S-FINANCIAL_ADVICE", "17": "B-HARMFUL_CONTENT", "18": "I-HARMFUL_CONTENT", "19": "E-HARMFUL_CONTENT", "20": "S-HARMFUL_CONTENT", "21": "B-CONFIDENTIAL_DISCLOSURE", "22": "I-CONFIDENTIAL_DISCLOSURE", "23": "E-CONFIDENTIAL_DISCLOSURE", "24": "S-CONFIDENTIAL_DISCLOSURE" }, "initial_context_length": 4096, "initializer_range": 0.02, "intermediate_size": 640, "label2id": { "B-COMPETITOR_MENTION": 1, "B-CONFIDENTIAL_DISCLOSURE": 21, "B-FINANCIAL_ADVICE": 13, "B-HARMFUL_CONTENT": 17, "B-PRIVACY_PII": 5, "B-PROFANITY": 9, "E-COMPETITOR_MENTION": 3, "E-CONFIDENTIAL_DISCLOSURE": 23, "E-FINANCIAL_ADVICE": 15, "E-HARMFUL_CONTENT": 19, "E-PRIVACY_PII": 7, "E-PROFANITY": 11, "I-COMPETITOR_MENTION": 2, "I-CONFIDENTIAL_DISCLOSURE": 22, "I-FINANCIAL_ADVICE": 14, "I-HARMFUL_CONTENT": 18, "I-PRIVACY_PII": 6, "I-PROFANITY": 10, "O": 0, "S-COMPETITOR_MENTION": 4, "S-CONFIDENTIAL_DISCLOSURE": 24, "S-FINANCIAL_ADVICE": 16, "S-HARMFUL_CONTENT": 20, "S-PRIVACY_PII": 8, "S-PROFANITY": 12 }, "max_position_embeddings": 131072, "model_type": "openai_privacy_filter", "num_attention_heads": 14, "num_experts_per_tok": 4, "num_hidden_layers": 8, "num_key_value_heads": 2, "num_local_experts": 128, "output_router_logits": false, "pad_token_id": 199999, "rms_norm_eps": 1e-05, "rope_parameters": { "beta_fast": 32.0, "beta_slow": 1.0, "factor": 32.0, "original_max_position_embeddings": 4096, "rope_theta": 150000.0, "rope_type": "yarn", "truncate": false }, "router_aux_loss_coef": 0.001, "sliding_window": 128, "tie_word_embeddings": false, "transformers.js_config": { "use_external_data_format": { "model": 1, "model.onnx": 3, "model_fp16.onnx": 2 } }, "transformers_version": "5.8.0.dev0", "use_cache": false, "vocab_size": 200064 }