Lacax commited on
Commit
50505f3
·
verified ·
1 Parent(s): 3d0158a

Upload model

Browse files
Files changed (3) hide show
  1. README.md +3 -0
  2. adapter_config.json +13 -9
  3. adapter_model.safetensors +2 -2
README.md CHANGED
@@ -1,4 +1,7 @@
1
  ---
 
 
 
2
  library_name: transformers
3
  pipeline_tag: image-text-to-text
4
  tags:
 
1
  ---
2
+ {}
3
+ ---
4
+ ---
5
  library_name: transformers
6
  pipeline_tag: image-text-to-text
7
  tags:
adapter_config.json CHANGED
@@ -4,8 +4,7 @@
4
  "arrow_config": null,
5
  "auto_mapping": {
6
  "base_model_class": "DeepseekOCR2ForCausalLM",
7
- "parent_library": "transformers_modules.deepseek_ocr2.modeling_deepseekocr2",
8
- "unsloth_fixed": true
9
  },
10
  "base_model_name_or_path": "./deepseek_ocr2",
11
  "bias": "none",
@@ -20,9 +19,9 @@
20
  "layers_pattern": null,
21
  "layers_to_transform": null,
22
  "loftq_config": {},
23
- "lora_alpha": 32,
24
  "lora_bias": false,
25
- "lora_dropout": 0,
26
  "megatron_config": null,
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
@@ -32,12 +31,17 @@
32
  "r": 32,
33
  "rank_pattern": {},
34
  "revision": null,
35
- "target_modules": "(?:.*?(?:vision|image|visual|patch|language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:q_proj|k_proj|v_proj|o_proj|gate_proj|up_proj|down_proj|qkv|proj|lin1|lin2).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:q_proj|k_proj|v_proj|o_proj|gate_proj|up_proj|down_proj|qkv|proj|lin1|lin2)))",
36
- "target_parameters": [
37
- "mlp.experts.gate_up_proj",
38
- "mlp.experts.down_proj"
 
 
 
 
39
  ],
40
- "task_type": "CAUSAL_LM",
 
41
  "trainable_token_indices": null,
42
  "use_dora": false,
43
  "use_qalora": false,
 
4
  "arrow_config": null,
5
  "auto_mapping": {
6
  "base_model_class": "DeepseekOCR2ForCausalLM",
7
+ "parent_library": "transformers_modules.deepseek_ocr2.modeling_deepseekocr2"
 
8
  },
9
  "base_model_name_or_path": "./deepseek_ocr2",
10
  "bias": "none",
 
19
  "layers_pattern": null,
20
  "layers_to_transform": null,
21
  "loftq_config": {},
22
+ "lora_alpha": 64,
23
  "lora_bias": false,
24
+ "lora_dropout": 0.05,
25
  "megatron_config": null,
26
  "megatron_core": "megatron.core",
27
  "modules_to_save": null,
 
31
  "r": 32,
32
  "rank_pattern": {},
33
  "revision": null,
34
+ "target_modules": [
35
+ "up_proj",
36
+ "o_proj",
37
+ "gate_proj",
38
+ "k_proj",
39
+ "down_proj",
40
+ "v_proj",
41
+ "q_proj"
42
  ],
43
+ "target_parameters": null,
44
+ "task_type": null,
45
  "trainable_token_indices": null,
46
  "use_dora": false,
47
  "use_qalora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44827d1a99a988512419a1a29e2aae9a17cfbe121938595950c0144c4b03acba
3
- size 18863392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3da86e8684d15d20b536dbe1a8cc81d8b00238dcd0952b3a4dd8fcc8c3dcf3b
3
+ size 691141456