rishi38 commited on
Commit
feb5e7a
·
verified ·
1 Parent(s): 2ccffe9

Training in progress, step 10

Browse files
README.md CHANGED
@@ -1,18 +1,18 @@
1
  ---
2
- base_model: unsloth/Qwen3-4B-unsloth-bnb-4bit
3
  library_name: transformers
4
  model_name: smart-emergency-grpo
5
  tags:
6
  - generated_from_trainer
7
  - unsloth
8
- - grpo
9
  - trl
 
10
  licence: license
11
  ---
12
 
13
  # Model Card for smart-emergency-grpo
14
 
15
- This model is a fine-tuned version of [unsloth/Qwen3-4B-unsloth-bnb-4bit](https://huggingface.co/unsloth/Qwen3-4B-unsloth-bnb-4bit).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
 
1
  ---
2
+ base_model: unsloth/Qwen3-1.7B-unsloth-bnb-4bit
3
  library_name: transformers
4
  model_name: smart-emergency-grpo
5
  tags:
6
  - generated_from_trainer
7
  - unsloth
 
8
  - trl
9
+ - grpo
10
  licence: license
11
  ---
12
 
13
  # Model Card for smart-emergency-grpo
14
 
15
+ This model is a fine-tuned version of [unsloth/Qwen3-1.7B-unsloth-bnb-4bit](https://huggingface.co/unsloth/Qwen3-1.7B-unsloth-bnb-4bit).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
adapter_config.json CHANGED
@@ -7,7 +7,7 @@
7
  "parent_library": "transformers.models.qwen3.modeling_qwen3",
8
  "unsloth_fixed": true
9
  },
10
- "base_model_name_or_path": "unsloth/Qwen3-4B-unsloth-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
@@ -34,13 +34,13 @@
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
37
- "up_proj",
38
- "down_proj",
39
- "gate_proj",
40
  "v_proj",
 
41
  "k_proj",
 
 
42
  "q_proj",
43
- "o_proj"
44
  ],
45
  "target_parameters": null,
46
  "task_type": "CAUSAL_LM",
 
7
  "parent_library": "transformers.models.qwen3.modeling_qwen3",
8
  "unsloth_fixed": true
9
  },
10
+ "base_model_name_or_path": "unsloth/Qwen3-1.7B-unsloth-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
  "ensure_weight_tying": false,
 
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
 
 
 
37
  "v_proj",
38
+ "o_proj",
39
  "k_proj",
40
+ "gate_proj",
41
+ "down_proj",
42
  "q_proj",
43
+ "up_proj"
44
  ],
45
  "target_parameters": null,
46
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e66882ce5e8ea97c7d9bb909d1348ae5c1b776a897b1b224bb2317adeea6f729
3
- size 132187888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:932b61a20c5b78ee61c932e84155a3c6bf1fea8a48a19c2857501b2471aacc34
3
+ size 69782384
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e3891a59089fdc7b05df276eddb9638706c9215c28a4e8f5db6693c2217a2ea
3
  size 7377
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b96489e9994db0566c3f6f6d74088be1436a13f4cc4986b0ecdd1bed3fbc3f2
3
  size 7377