jeromeramos commited on
Commit
e461471
·
verified ·
1 Parent(s): 49049db

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "Sim4Rec/inter-play-sim-assistant-sft",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -18,7 +17,7 @@
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
21
- "pad_token_id": 128001,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
@@ -31,7 +30,7 @@
31
  "rope_theta": 500000.0,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.48.2",
35
- "use_cache": true,
36
  "vocab_size": 128320
37
  }
 
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
 
17
  "num_attention_heads": 32,
18
  "num_hidden_layers": 32,
19
  "num_key_value_heads": 8,
20
+ "pad_token_id": 128004,
21
  "pretraining_tp": 1,
22
  "rms_norm_eps": 1e-05,
23
  "rope_scaling": {
 
30
  "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.51.3",
34
+ "use_cache": false,
35
  "vocab_size": 128320
36
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1dca7caf67b0af32b1aa5c9c52ad02de5b54e84a8ec038bd74e3bc91b18d8c7
3
  size 4977222960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153e9b2f9968cee7c24500b9db39043ef6774e83253ae1c5e9f0334989f34523
3
  size 4977222960
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f09aa5fdc5b3591ad4b03edba26dce3703417eba2a78ff25e3f469b3664a254
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0840f1cb9d760a5fe7cdf22d38bd1850143ce43ad67a9f237e72e62974a568c3
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da6625e7a5cfcdd77627f26512cf498048fc24b6547e71e3ef4c3a71e41141d9
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e0cbc723f505f812a453e788f1e4eca733497803384056017b633870d1cdf1e
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6115a28a38e67e9c28cd5db2742d2794509bfac5395d3a4d4f58a995d926ed67
3
  size 1168663096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:246dc7b1f34895bf90a6b0804a19d4ded9e8e98d711a3372086d2271bee3028b
3
  size 1168663096
runs/May07_15-46-50_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-84pj9pb/events.out.tfevents.1746632959.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-84pj9pb.15363.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c1159dea1e8aef4ef9c1581b1bd648f29fa0c5dba4e7f59f1d74ab5d8b845cb
3
+ size 6312
runs/May07_18-58-10_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-7bjk8jd/events.out.tfevents.1746644432.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-7bjk8jd.16849.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:367914b0c34f4d4fe626a3c404b5dad5a5c625a85d972bc89f18d111688a09fe
3
+ size 203122
special_tokens_map.json CHANGED
@@ -2,9 +2,16 @@
2
  "additional_special_tokens": [
3
  "<response>",
4
  "</response>",
5
- "<answer>",
6
- "</answer>",
7
- "<inquire>"
 
 
 
 
 
 
 
8
  ],
9
  "bos_token": {
10
  "content": "<|im_start|>",
@@ -21,7 +28,7 @@
21
  "single_word": false
22
  },
23
  "pad_token": {
24
- "content": "<|end_of_text|>",
25
  "lstrip": false,
26
  "normalized": false,
27
  "rstrip": false,
 
2
  "additional_special_tokens": [
3
  "<response>",
4
  "</response>",
5
+ "<action>",
6
+ "</action>",
7
+ "<inquire>",
8
+ "<recommend>",
9
+ "<greeting>",
10
+ "<disclose-goal>",
11
+ "<feedback>",
12
+ "<accept>",
13
+ "<movie_title>",
14
+ "</movie_title>"
15
  ],
16
  "bos_token": {
17
  "content": "<|im_start|>",
 
28
  "single_word": false
29
  },
30
  "pad_token": {
31
+ "content": "<|finetune_right_pad_id|>",
32
  "lstrip": false,
33
  "normalized": false,
34
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5ea5afcc70a5f73f9b545a5940b211fd23e2acd4d895a3ebc3144ca348a4633
3
- size 17211228
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1333f4772561fe7012579974661e97e8d5105c4a40fcaeca1fcaa35e2f1c72f7
3
+ size 17212548
tokenizer_config.json CHANGED
@@ -2081,7 +2081,7 @@
2081
  "special": true
2082
  },
2083
  "128260": {
2084
- "content": "<answer>",
2085
  "lstrip": false,
2086
  "normalized": false,
2087
  "rstrip": false,
@@ -2089,7 +2089,7 @@
2089
  "special": true
2090
  },
2091
  "128261": {
2092
- "content": "</answer>",
2093
  "lstrip": false,
2094
  "normalized": false,
2095
  "rstrip": false,
@@ -2103,29 +2103,88 @@
2103
  "rstrip": false,
2104
  "single_word": false,
2105
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2106
  }
2107
  },
2108
  "additional_special_tokens": [
2109
  "<response>",
2110
  "</response>",
2111
- "<answer>",
2112
- "</answer>",
2113
- "<inquire>"
 
 
 
 
 
 
 
2114
  ],
2115
  "bos_token": "<|im_start|>",
2116
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2117
  "clean_up_tokenization_spaces": true,
2118
  "eos_token": "<|im_end|>",
2119
  "extra_special_tokens": {},
2120
- "max_length": 2048,
2121
  "model_input_names": [
2122
  "input_ids",
2123
  "attention_mask"
2124
  ],
2125
  "model_max_length": 131072,
2126
- "pad_token": "<|end_of_text|>",
2127
- "stride": 0,
2128
- "tokenizer_class": "PreTrainedTokenizerFast",
2129
- "truncation_side": "right",
2130
- "truncation_strategy": "longest_first"
2131
  }
 
2081
  "special": true
2082
  },
2083
  "128260": {
2084
+ "content": "<action>",
2085
  "lstrip": false,
2086
  "normalized": false,
2087
  "rstrip": false,
 
2089
  "special": true
2090
  },
2091
  "128261": {
2092
+ "content": "</action>",
2093
  "lstrip": false,
2094
  "normalized": false,
2095
  "rstrip": false,
 
2103
  "rstrip": false,
2104
  "single_word": false,
2105
  "special": true
2106
+ },
2107
+ "128263": {
2108
+ "content": "<recommend>",
2109
+ "lstrip": false,
2110
+ "normalized": false,
2111
+ "rstrip": false,
2112
+ "single_word": false,
2113
+ "special": true
2114
+ },
2115
+ "128264": {
2116
+ "content": "<greeting>",
2117
+ "lstrip": false,
2118
+ "normalized": false,
2119
+ "rstrip": false,
2120
+ "single_word": false,
2121
+ "special": true
2122
+ },
2123
+ "128265": {
2124
+ "content": "<disclose-goal>",
2125
+ "lstrip": false,
2126
+ "normalized": false,
2127
+ "rstrip": false,
2128
+ "single_word": false,
2129
+ "special": true
2130
+ },
2131
+ "128266": {
2132
+ "content": "<feedback>",
2133
+ "lstrip": false,
2134
+ "normalized": false,
2135
+ "rstrip": false,
2136
+ "single_word": false,
2137
+ "special": true
2138
+ },
2139
+ "128267": {
2140
+ "content": "<accept>",
2141
+ "lstrip": false,
2142
+ "normalized": false,
2143
+ "rstrip": false,
2144
+ "single_word": false,
2145
+ "special": true
2146
+ },
2147
+ "128268": {
2148
+ "content": "<movie_title>",
2149
+ "lstrip": false,
2150
+ "normalized": false,
2151
+ "rstrip": false,
2152
+ "single_word": false,
2153
+ "special": true
2154
+ },
2155
+ "128269": {
2156
+ "content": "</movie_title>",
2157
+ "lstrip": false,
2158
+ "normalized": false,
2159
+ "rstrip": false,
2160
+ "single_word": false,
2161
+ "special": true
2162
  }
2163
  },
2164
  "additional_special_tokens": [
2165
  "<response>",
2166
  "</response>",
2167
+ "<action>",
2168
+ "</action>",
2169
+ "<inquire>",
2170
+ "<recommend>",
2171
+ "<greeting>",
2172
+ "<disclose-goal>",
2173
+ "<feedback>",
2174
+ "<accept>",
2175
+ "<movie_title>",
2176
+ "</movie_title>"
2177
  ],
2178
  "bos_token": "<|im_start|>",
2179
  "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2180
  "clean_up_tokenization_spaces": true,
2181
  "eos_token": "<|im_end|>",
2182
  "extra_special_tokens": {},
 
2183
  "model_input_names": [
2184
  "input_ids",
2185
  "attention_mask"
2186
  ],
2187
  "model_max_length": 131072,
2188
+ "pad_token": "<|finetune_right_pad_id|>",
2189
+ "tokenizer_class": "PreTrainedTokenizer"
 
 
 
2190
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e31315ddeb7aa7465f2bd16d20b45c43e929c9c6b6b9f79d4646ff479ab0b680
3
  size 7800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:395fb48e7de57f282dbfbd55c1cdbbb8e8a6be71123f5d59c2730f53296e6c0e
3
  size 7800