akhiilll commited on
Commit
9d99dff
·
verified ·
1 Parent(s): d0380fe

GRPO LoRA (sft=1000, grpo=200)

Browse files
adapter_config.json CHANGED
@@ -7,7 +7,7 @@
7
  "parent_library": "transformers.models.qwen2.modeling_qwen2",
8
  "unsloth_fixed": true
9
  },
10
- "base_model_name_or_path": "unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
@@ -34,13 +34,13 @@
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
37
- "down_proj",
38
- "k_proj",
39
- "o_proj",
40
  "up_proj",
 
41
  "v_proj",
42
- "gate_proj",
43
- "q_proj"
 
 
44
  ],
45
  "target_parameters": null,
46
  "task_type": "CAUSAL_LM",
 
7
  "parent_library": "transformers.models.qwen2.modeling_qwen2",
8
  "unsloth_fixed": true
9
  },
10
+ "base_model_name_or_path": "unsloth/qwen2.5-coder-7b-instruct-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
 
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
 
 
 
37
  "up_proj",
38
+ "k_proj",
39
  "v_proj",
40
+ "down_proj",
41
+ "q_proj",
42
+ "o_proj",
43
+ "gate_proj"
44
  ],
45
  "target_parameters": null,
46
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55e6151b16e8af4a8aaf0b7cfdac763ffd8825c85397ece9e1b582f64e6b5fba
3
- size 119801528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0b4c9947fcc438880784fc8fe35481f75c42270a97efedac74afc7831d39ae0
3
+ size 161533192
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af5891a15588546db1ac7f2baf8fa94835a51a85c032c39793a55bb048b47446
3
- size 11422523
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea43b288542655d72d632195ab9b58ca2cd9532c292bf6667827ce899ad196bc
3
+ size 11422082
tokenizer_config.json CHANGED
@@ -5,7 +5,9 @@
5
  "clean_up_tokenization_spaces": false,
6
  "eos_token": "<|im_end|>",
7
  "errors": "replace",
 
8
  "is_local": false,
 
9
  "model_max_length": 32768,
10
  "pad_token": "<|PAD_TOKEN|>",
11
  "padding_side": "right",
 
5
  "clean_up_tokenization_spaces": false,
6
  "eos_token": "<|im_end|>",
7
  "errors": "replace",
8
+ "extra_special_tokens": [],
9
  "is_local": false,
10
+ "local_files_only": false,
11
  "model_max_length": 32768,
12
  "pad_token": "<|PAD_TOKEN|>",
13
  "padding_side": "right",