interpgpt-adhd-23M / configuration_interpgpt.py
connaaa's picture
fix: avoid duplicate pad_token_id kwarg on config reload
a0c5694 verified
"""
HuggingFace PretrainedConfig for InterpGPT / TaskGPT.
Mirrors gpt_model.GPTConfig but subclasses transformers.PretrainedConfig
so `AutoConfig` / `AutoModel.from_pretrained(..., trust_remote_code=True)` work.
"""
from transformers import PretrainedConfig
class InterpGPTConfig(PretrainedConfig):
model_type = "interpgpt"
def __init__(
self,
vocab_size: int = 8192,
max_seq_len: int = 512,
n_layers: int = 6,
n_heads: int = 8,
d_model: int = 512,
d_ff: int = 2048,
dropout: float = 0.1,
pad_id: int = 0,
bias: bool = False,
variant: str = "standard",
**kwargs,
):
self.vocab_size = vocab_size
self.max_seq_len = max_seq_len
self.n_layers = n_layers
self.n_heads = n_heads
self.d_model = d_model
self.d_ff = d_ff
self.dropout = dropout
self.pad_id = pad_id
self.bias = bias
self.variant = variant
kwargs.pop("pad_token_id", None)
super().__init__(pad_token_id=pad_id, **kwargs)