TranTruongMMCII commited on
Commit
3c9dad1
·
verified ·
1 Parent(s): dba904e

Upload 4 files

Browse files
checkpoint-best/baseline/config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <<<<<<< HEAD
2
+ version https://git-lfs.github.com/spec/v1
3
+ oid sha256:242631d24516c99f8ee9714306ff67830f38e99553d0e78101f812f7d7c33737
4
+ size 880
5
+ =======
6
+ {
7
+ "_num_labels": 2,
8
+ "activation_function": "gelu_new",
9
+ "add_cross_attention": false,
10
+ "architectures": [
11
+ "GPT2LMHeadModel"
12
+ ],
13
+ "attn_pdrop": 0.1,
14
+ "bos_token_id": 0,
15
+ "dtype": "float32",
16
+ "embd_pdrop": 0.1,
17
+ "eos_token_id": 2,
18
+ "initializer_range": 0.02,
19
+ "layer_norm_epsilon": 1e-05,
20
+ "model_type": "gpt2",
21
+ "n_ctx": 1024,
22
+ "n_embd": 768,
23
+ "n_head": 12,
24
+ "n_inner": null,
25
+ "n_layer": 12,
26
+ "n_positions": 1024,
27
+ "output_past": true,
28
+ "pad_token_id": 1,
29
+ "reorder_and_upcast_attn": false,
30
+ "resid_pdrop": 0.1,
31
+ "scale_attn_by_inverse_layer_idx": false,
32
+ "scale_attn_weights": true,
33
+ "summary_activation": null,
34
+ "summary_first_dropout": 0.1,
35
+ "summary_proj_to_labels": true,
36
+ "summary_type": "cls_index",
37
+ "summary_use_proj": true,
38
+ "tie_word_embeddings": true,
39
+ "transformers_version": "5.0.0",
40
+ "use_cache": true,
41
+ "vocab_size": 50007
42
+ }
43
+ >>>>>>> 91e23bbded3eb35d26872429e85ce50dbc971d12
checkpoint-best/baseline/generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <<<<<<< HEAD
2
+ version https://git-lfs.github.com/spec/v1
3
+ oid sha256:f2c690d1ae1ac19a8895c1d6905eed6337d0276a1611afca0a6fdd83cef49da4
4
+ size 131
5
+ =======
6
+ {
7
+ "_from_model_config": true,
8
+ "bos_token_id": 0,
9
+ "eos_token_id": 2,
10
+ "pad_token_id": 1,
11
+ "transformers_version": "5.0.0"
12
+ }
13
+ >>>>>>> 91e23bbded3eb35d26872429e85ce50dbc971d12
checkpoint-best/baseline/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-best/baseline/tokenizer_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <<<<<<< HEAD
2
+ version https://git-lfs.github.com/spec/v1
3
+ oid sha256:c4654ce1e1fff683f3851e0e83aa1a58cd33e2716ea2d0a0094bf11e2880a7e8
4
+ size 449
5
+ =======
6
+ {
7
+ "add_prefix_space": false,
8
+ "backend": "tokenizers",
9
+ "bos_token": "<s>",
10
+ "eos_token": "</s>",
11
+ "errors": "replace",
12
+ "extra_special_tokens": [
13
+ "<RET>",
14
+ "</RET>",
15
+ "<CTX>",
16
+ "</CTX>",
17
+ "<GEN>"
18
+ ],
19
+ "full_tokenizer_file": null,
20
+ "is_local": false,
21
+ "model_max_length": 1000000000000000019884624838656,
22
+ "pad_token": "<pad>",
23
+ "sep_token": "<EOL>",
24
+ "tokenizer_class": "GPT2Tokenizer",
25
+ "unk_token": "<|UNKNOWN|>"
26
+ }
27
+ >>>>>>> 91e23bbded3eb35d26872429e85ce50dbc971d12