jacobfa1 commited on
Commit
0730d49
·
verified ·
1 Parent(s): 65efe25

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +65 -0
  2. model.pt +3 -0
config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "eqlm-small-140m",
3
+ "hf_config": {
4
+ "org": "SandyResearch",
5
+ "name": "eqlm-small-140m"
6
+ },
7
+ "block_size": 2048,
8
+ "n_embd": 1024,
9
+ "intermediate_size": 4096,
10
+ "num_attention_heads": 8,
11
+ "num_key_value_heads": 8,
12
+ "vocab_size": 32768,
13
+ "padding_multiple": 64,
14
+ "padded_vocab_size": 32768,
15
+ "rope_settings": {
16
+ "use_rope": true,
17
+ "rope_condense_ratio": 1,
18
+ "rope_base": 50000
19
+ },
20
+ "use_abacus": false,
21
+ "randomize_positions_from": null,
22
+ "block_class_name": "TransformerPreNormBlock",
23
+ "norm_class_name": "RMSNorm",
24
+ "attn_impl": "flash",
25
+ "norm_eps": 1e-05,
26
+ "mlp_class_name": "BaseMLP",
27
+ "nonlin_name": "ReLU2",
28
+ "bias": false,
29
+ "qk_bias": false,
30
+ "init_strategy": "scaled-zero",
31
+ "init_orthogonal": true,
32
+ "skip_initialization": false,
33
+ "mup_model_scaling_factor": 1,
34
+ "use_fused_head": "pytorch",
35
+ "debias_attention": false,
36
+ "center_attention": false,
37
+ "clip_qkv": null,
38
+ "qk_norm": true,
39
+ "logit_softcap": null,
40
+ "causal": true,
41
+ "activation_checkpoint_impl": "per-block",
42
+ "simple_ops": false,
43
+ "strategy": "ddp",
44
+ "n_backbone_layers": 7,
45
+ "n_fp_blocks": 1,
46
+ "tie_embeddings": true,
47
+ "solver": "anderson",
48
+ "max_iter": 64,
49
+ "min_iter": 6,
50
+ "tol": 0.0003,
51
+ "anderson_m": 5,
52
+ "anderson_beta": 1.0,
53
+ "backward_type": "onestep",
54
+ "backward_max_iter": 64,
55
+ "backward_min_iter": 6,
56
+ "backward_tol": 0.0003,
57
+ "adjoint_grad_clip": null,
58
+ "layer_scale_init": 0.75,
59
+ "gamma_max": 0.75,
60
+ "fp_lr_scale": 0.5,
61
+ "fp_wd": 0.1,
62
+ "recurrent_embedding_dimension": 1024,
63
+ "model_class_name": "EQLM",
64
+ "_class_name": "EQLMConfig"
65
+ }
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea49a8d1aa841565c20bf2177e63843098071bbc3b7155b47b0646b40bfb2e43
3
+ size 538020542