| from transformers.configuration_utils import PretrainedConfig |
| from typing import Any |
|
|
| class AtomformerConfig(PretrainedConfig): |
| r""" |
| Configuration of a :class:`~transform:class:`~transformers.AtomformerModel`. |
| |
| It is used to instantiate an Atomformer model according to the specified arguments. |
| """ |
|
|
| model_type = "atomformer" |
|
|
| def __init__( |
| self, |
| vocab_size: int = 123, |
| dim: int = 768, |
| num_heads: int = 32, |
| depth: int = 12, |
| mlp_ratio: int = 1, |
| k: int = 128, |
| dropout: float = 0.0, |
| mask_token_id: int = 0, |
| pad_token_id: int = 119, |
| bos_token_id: int = 120, |
| eos_token_id: int = 121, |
| cls_token_id: int = 122, |
| **kwargs: Any, |
| ) -> None: |
| super().__init__(**kwargs) |
| self.vocab_size = vocab_size |
| self.dim = dim |
| self.num_heads = num_heads |
| self.depth = depth |
| self.mlp_ratio = mlp_ratio |
| self.k = k |
|
|
| self.dropout = dropout |
| self.mask_token_id = mask_token_id |
| self.pad_token_id = pad_token_id |
| self.bos_token_id = bos_token_id |
| self.eos_token_id = eos_token_id |
| self.cls_token_id = cls_token_id |