| { |
| "model_type": "minitransformer", |
| "_name_or_path": "Transformer_500M", |
| "architectures": ["MiniTransformer"], |
| "dim": 768, |
| "num_heads": 24, |
| "num_layers": 27, |
| "seq_len": 8192, |
| "window_size": 8192, |
| "vocab_size": 200064, |
| "mlp_scale": 4, |
| "bias": false, |
| "dropout": 0.0, |
| "num_epochs": 1, |
| "global_bsz": 524288, |
| "bsz": 1, |
| "warmup_steps": 1907, |
| "eval_period": 50, |
| "save_period": 500, |
| "max_lr": 3.0e-4, |
| "min_lr": 3.0e-5, |
| "max_norm": 1.0, |
| "dilation": 1, |
| "fsdp": false, |
| "ddp": true, |
| "mixed_precision": true, |
| "torch_dtype": "bfloat16", |
| "cpu_offload": false, |
| "sharding_strategy": "full_shard", |
| "state_dict_type": "full", |
| "auto_wrap_policy": "partial", |
| "backward_prefetch": "backward_pre", |
| "forward_prefetch": false, |
| "sync_module_states": true, |
| "use_orig_params": true, |
| "device_id": null, |
| "precision": { |
| "param": "bfloat16", |
| "reduce": "bfloat16", |
| "buffer": "bfloat16" |
| }, |
| "fsdp_modules": [ |
| "AttentionLayer" |
| ], |
| "use_activation_checkpointing": true, |
| "softcap": 50.0, |
| "theta": 10000.0, |
| "torch_compile": true |
| } |