Lolol857 commited on
Commit
76f0fee
·
verified ·
1 Parent(s): 3655d63

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -34,12 +34,12 @@
34
  "revision": null,
35
  "target_modules": [
36
  "v_proj",
37
- "gate_proj",
38
  "down_proj",
39
  "q_proj",
 
40
  "up_proj",
41
  "o_proj",
42
- "k_proj"
43
  ],
44
  "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
 
34
  "revision": null,
35
  "target_modules": [
36
  "v_proj",
 
37
  "down_proj",
38
  "q_proj",
39
+ "k_proj",
40
  "up_proj",
41
  "o_proj",
42
+ "gate_proj"
43
  ],
44
  "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f6ffbbc3ef1f6459a42262c20daa57644415f98821543c7715481de3e049f96
3
- size 71320216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:004a710a01739ae3c15a50d64e44652fec98d5ecb61417891e3231a60dc3e913
3
+ size 69157536
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 199999,
4
+ "eos_token_id": [
5
+ 200020
6
+ ],
7
+ "max_length": 131072,
8
+ "pad_token_id": 200029,
9
+ "transformers_version": "5.5.0"
10
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbcf8ec8306b4f270411b96a1434505a36b36c0a56d79284f15e47d93ed13462
3
  size 5649
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4aa2e3325989e8434fe3c0f7564866f88b52c2009f394bc964c603eedb7a1a38
3
  size 5649