williamium commited on
Commit
c3901f3
·
verified ·
1 Parent(s): d34fa06

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_0.pt +3 -0
  3. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_1.pt +3 -0
  4. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_2.pt +3 -0
  5. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_3.pt +3 -0
  6. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_4.pt +3 -0
  7. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_5.pt +3 -0
  8. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_6.pt +3 -0
  9. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_7.pt +3 -0
  10. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/fsdp_config.json +4 -0
  11. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/added_tokens.json +24 -0
  12. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/chat_template.jinja +7 -0
  13. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/config.json +132 -0
  14. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/generation_config.json +12 -0
  15. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/merges.txt +0 -0
  16. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/preprocessor_config.json +39 -0
  17. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/special_tokens_map.json +31 -0
  18. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/tokenizer.json +3 -0
  19. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/tokenizer_config.json +208 -0
  20. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/video_preprocessor_config.json +43 -0
  21. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/vocab.json +0 -0
  22. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_0.pt +3 -0
  23. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_1.pt +3 -0
  24. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_2.pt +3 -0
  25. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_3.pt +3 -0
  26. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_4.pt +3 -0
  27. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_5.pt +3 -0
  28. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_6.pt +3 -0
  29. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_7.pt +3 -0
  30. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_0.pt +3 -0
  31. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_1.pt +3 -0
  32. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_2.pt +3 -0
  33. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_3.pt +3 -0
  34. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_4.pt +3 -0
  35. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_5.pt +3 -0
  36. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_6.pt +3 -0
  37. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_7.pt +3 -0
  38. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/data.pt +3 -0
  39. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_0.pt +3 -0
  40. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_1.pt +3 -0
  41. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_2.pt +3 -0
  42. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_3.pt +3 -0
  43. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_4.pt +3 -0
  44. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_5.pt +3 -0
  45. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_6.pt +3 -0
  46. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_7.pt +3 -0
  47. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/fsdp_config.json +4 -0
  48. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/added_tokens.json +24 -0
  49. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/chat_template.jinja +7 -0
  50. grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/config.json +132 -0
.gitattributes CHANGED
@@ -35,3 +35,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  llava_vqa_judge_grpo_local_qwen3/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  llava_vqa_judge_grpo_local_qwen3_new_1epoch/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  llava_vqa_judge_grpo_local_qwen3/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  llava_vqa_judge_grpo_local_qwen3_new_1epoch/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_30/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
41
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_40/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_48/actor/huggingface/tokenizer.json filter=lfs diff=lfs merge=lfs -text
43
+ grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1_2026-03-31_18-23-55.log filter=lfs diff=lfs merge=lfs -text
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b16d86252a04c84cea641a4f5a57b9abf26a189c57145c92c5d8c9c21608a05
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b12150881d1c4a3c4a212bdd503c8ae6bb6b015a9b0ec3b94b3346fae1aecb2a
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aab917eca73cbbe9dc02e9c86f7f60c96163cd55ef8c9ad57014404468f10b7a
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fdbcf04e75d1b3df41f80930ea503336d8af3e8030050b94efd6f46a1f6381d
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c77b8a5759d8117bb602706949f672bb63b7a88de47df3c0067c811a5e4d118a
3
+ size 15205
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b9cc184d2be40f9ab11dbaf5898491da1b28daec26f6dcb596a4516340901db
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfe8162fe2d1e28b3e1a61d33f6eef535213e447416566a131933bdf63d9f8d5
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/extra_state_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:befb575db58c6b645eb271b4df127ad5dfc6df8b67546cb1ae5b97fa0a281359
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/fsdp_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "FSDP_version": 1,
3
+ "world_size": 8
4
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/chat_template.jinja ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system
2
+ You are a helpful assistant.<|im_end|>
3
+ {% endif %}<|im_start|>{{ message['role'] }}
4
+ {% if message['content'] is string %}{{ message['content'] }}<|im_end|>
5
+ {% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>
6
+ {% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant
7
+ {% endif %}
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/config.json ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2_5_VLForConditionalGeneration"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "dtype": "float32",
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "image_token_id": 151655,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 18944,
14
+ "max_position_embeddings": 128000,
15
+ "max_window_layers": 28,
16
+ "model_type": "qwen2_5_vl",
17
+ "num_attention_heads": 28,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 4,
20
+ "rms_norm_eps": 1e-06,
21
+ "rope_scaling": {
22
+ "mrope_section": [
23
+ 16,
24
+ 24,
25
+ 24
26
+ ],
27
+ "rope_type": "default",
28
+ "type": "default"
29
+ },
30
+ "rope_theta": 1000000.0,
31
+ "sliding_window": 32768,
32
+ "text_config": {
33
+ "_name_or_path": "../ckpts/Qwen2.5-VL-7B-Instruct",
34
+ "architectures": [
35
+ "Qwen2_5_VLForConditionalGeneration"
36
+ ],
37
+ "attention_dropout": 0.0,
38
+ "dtype": "float32",
39
+ "eos_token_id": 151645,
40
+ "hidden_act": "silu",
41
+ "hidden_size": 3584,
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 18944,
44
+ "layer_types": [
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention",
53
+ "full_attention",
54
+ "full_attention",
55
+ "full_attention",
56
+ "full_attention",
57
+ "full_attention",
58
+ "full_attention",
59
+ "full_attention",
60
+ "full_attention",
61
+ "full_attention",
62
+ "full_attention",
63
+ "full_attention",
64
+ "full_attention",
65
+ "full_attention",
66
+ "full_attention",
67
+ "full_attention",
68
+ "full_attention",
69
+ "full_attention",
70
+ "full_attention",
71
+ "full_attention",
72
+ "full_attention"
73
+ ],
74
+ "max_position_embeddings": 128000,
75
+ "max_window_layers": 28,
76
+ "model_type": "qwen2_5_vl_text",
77
+ "num_attention_heads": 28,
78
+ "num_hidden_layers": 28,
79
+ "num_key_value_heads": 4,
80
+ "pad_token_id": 151643,
81
+ "rms_norm_eps": 1e-06,
82
+ "rope_scaling": {
83
+ "mrope_section": [
84
+ 16,
85
+ 24,
86
+ 24
87
+ ],
88
+ "rope_type": "default",
89
+ "type": "default"
90
+ },
91
+ "rope_theta": 1000000.0,
92
+ "sliding_window": null,
93
+ "use_cache": true,
94
+ "use_sliding_window": false,
95
+ "vision_token_id": 151654,
96
+ "vocab_size": 152064
97
+ },
98
+ "tie_word_embeddings": false,
99
+ "transformers_version": "4.57.1",
100
+ "use_cache": true,
101
+ "use_sliding_window": false,
102
+ "video_token_id": 151656,
103
+ "vision_config": {
104
+ "depth": 32,
105
+ "dtype": "float32",
106
+ "fullatt_block_indexes": [
107
+ 7,
108
+ 15,
109
+ 23,
110
+ 31
111
+ ],
112
+ "hidden_act": "silu",
113
+ "hidden_size": 1280,
114
+ "in_channels": 3,
115
+ "in_chans": 3,
116
+ "initializer_range": 0.02,
117
+ "intermediate_size": 3420,
118
+ "model_type": "qwen2_5_vl",
119
+ "num_heads": 16,
120
+ "out_hidden_size": 3584,
121
+ "patch_size": 14,
122
+ "spatial_merge_size": 2,
123
+ "spatial_patch_size": 14,
124
+ "temporal_patch_size": 2,
125
+ "tokens_per_second": 2,
126
+ "window_size": 112
127
+ },
128
+ "vision_end_token_id": 151653,
129
+ "vision_start_token_id": 151652,
130
+ "vision_token_id": 151654,
131
+ "vocab_size": 152064
132
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 1e-06,
11
+ "transformers_version": "4.57.1"
12
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/preprocessor_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": true,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "do_center_crop": null,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": null,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_mean": [
14
+ 0.48145466,
15
+ 0.4578275,
16
+ 0.40821073
17
+ ],
18
+ "image_processor_type": "Qwen2VLImageProcessorFast",
19
+ "image_std": [
20
+ 0.26862954,
21
+ 0.26130258,
22
+ 0.27577711
23
+ ],
24
+ "input_data_format": null,
25
+ "max_pixels": 12845056,
26
+ "merge_size": 2,
27
+ "min_pixels": 3136,
28
+ "pad_size": null,
29
+ "patch_size": 14,
30
+ "processor_class": "Qwen2_5_VLProcessor",
31
+ "resample": 3,
32
+ "rescale_factor": 0.00392156862745098,
33
+ "return_tensors": null,
34
+ "size": {
35
+ "longest_edge": 12845056,
36
+ "shortest_edge": 3136
37
+ },
38
+ "temporal_patch_size": 2
39
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "model_max_length": 131072,
203
+ "pad_token": "<|endoftext|>",
204
+ "processor_class": "Qwen2_5_VLProcessor",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/video_preprocessor_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": true,
5
+ "device": null,
6
+ "do_center_crop": null,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "do_sample_frames": false,
12
+ "fps": null,
13
+ "image_mean": [
14
+ 0.48145466,
15
+ 0.4578275,
16
+ 0.40821073
17
+ ],
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "input_data_format": null,
24
+ "max_frames": 768,
25
+ "max_pixels": 12845056,
26
+ "merge_size": 2,
27
+ "min_frames": 4,
28
+ "min_pixels": 3136,
29
+ "num_frames": null,
30
+ "pad_size": null,
31
+ "patch_size": 14,
32
+ "processor_class": "Qwen2_5_VLProcessor",
33
+ "resample": 3,
34
+ "rescale_factor": 0.00392156862745098,
35
+ "return_metadata": false,
36
+ "size": {
37
+ "longest_edge": 12845056,
38
+ "shortest_edge": 3136
39
+ },
40
+ "temporal_patch_size": 2,
41
+ "video_metadata": null,
42
+ "video_processor_type": "Qwen2VLVideoProcessor"
43
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/huggingface/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fea3bebc626a4a321c35a328405f45940a789e7ba2a3bbdede1e1122dfa160d8
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:743d4672211f853367279067a68266e2406da1e4f5595eedde4629d62930e088
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49be464c16f889185e258e3223be4230e546db1a933bcfc240d1d50b2734038a
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deb95504f160b3a27781f54365ecbb5dc8ba0aa7cfce02663e3e5f8535b9ebf6
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53d96915ca2094912e26fe711aac11846592a2575a39faaaca70e783a8a109c0
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84449eaccbbbd9d0c1964e1bde8022d4036da1d02847f18737323fa70abbc7d2
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:526d0dbb04794f0d28da13fede71369b0f57d79ba6a20a2af33964ac62a4cd21
3
+ size 4146658699
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/model_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0643343f719f29bfb44575df7011ac437418eb56c279152902410933e26740d
3
+ size 4145343883
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de66d673e52dd90865f1ca539b957b15b36c4110d13e4caec7f030df57b849a6
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37a831c7dd0ac0a12aed03af9856fbc15dc4c1faaf1d3132c3475124e6357625
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61146fc533549efab8358dfd938b18d81ae5b0d8db77aa5bfc6e8af0ec6ff459
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1106d0f9aa4cc78f19d006f3b7bdc373698f0a8f9d8b065ea667d252fabd2946
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:339c6fecabc4710a351ec9a926d1f30d122e1b7c12e117ff69e294744be84952
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02895a5904001fe9f3b4c1a7cdf89034eea3f3a106e06bd5b3ab80ed135deb9e
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62d36304ac4f280a8229aef6e02ebf24ff594c05a9dbc25034eb43732f6ee857
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/actor/optim_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37fe4f7af1dc19db95ebc9e0cb4a5c3956a732c8b2ff28f185570134f101d506
3
+ size 8292331679
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_10/data.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9fd9f4dd1e1bac1f0caab81813e6f5e02b8141eef28574e7efcf9dfa077c0d3
3
+ size 7316
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfafebb094f5c699cbaf7fcab36c6f1237bd115a7fbecc497ff56bb2ba3ece13
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30dd178771a2aedac938ee7eb4668a425c9fd6dba24c03b5f3379d2fc1bc1d11
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15bad2d440fefdd55b126fcd8487ad2cfb24fbd5d9b8a20bde8524b951a77b40
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd493d754fe600a62c03d5432ee3bb94b522cc7fc6fcd9ec5b8ca49d352c64a2
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_4.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e75c85d3023ed95eb71f0ba1fbb294d4346508b2a80ab2246a03505eec87bde5
3
+ size 15205
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_5.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee186ca62005ac0f710b77eabf643516b284594af4183aadda5f9c54b2256345
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8280c5ff59fadd3d12bbdb130a58b4eebd7369ac0d287e112499a549dcad2040
3
+ size 15141
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/extra_state_world_size_8_rank_7.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccc6be9b7353f24a752fe76a5815bc7df1361fe0a81127f5f9c23add32110cf6
3
+ size 15077
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/fsdp_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "FSDP_version": 1,
3
+ "world_size": 8
4
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/chat_template.jinja ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system
2
+ You are a helpful assistant.<|im_end|>
3
+ {% endif %}<|im_start|>{{ message['role'] }}
4
+ {% if message['content'] is string %}{{ message['content'] }}<|im_end|>
5
+ {% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>
6
+ {% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant
7
+ {% endif %}
grpo_Qwen2.5-VL-7B-Instruct_tr1024_mini256_micro32_n8_lr5e-6_cosine_ep1/global_step_20/actor/huggingface/config.json ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2_5_VLForConditionalGeneration"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "dtype": "float32",
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "image_token_id": 151655,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 18944,
14
+ "max_position_embeddings": 128000,
15
+ "max_window_layers": 28,
16
+ "model_type": "qwen2_5_vl",
17
+ "num_attention_heads": 28,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 4,
20
+ "rms_norm_eps": 1e-06,
21
+ "rope_scaling": {
22
+ "mrope_section": [
23
+ 16,
24
+ 24,
25
+ 24
26
+ ],
27
+ "rope_type": "default",
28
+ "type": "default"
29
+ },
30
+ "rope_theta": 1000000.0,
31
+ "sliding_window": 32768,
32
+ "text_config": {
33
+ "_name_or_path": "../ckpts/Qwen2.5-VL-7B-Instruct",
34
+ "architectures": [
35
+ "Qwen2_5_VLForConditionalGeneration"
36
+ ],
37
+ "attention_dropout": 0.0,
38
+ "dtype": "float32",
39
+ "eos_token_id": 151645,
40
+ "hidden_act": "silu",
41
+ "hidden_size": 3584,
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 18944,
44
+ "layer_types": [
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention",
53
+ "full_attention",
54
+ "full_attention",
55
+ "full_attention",
56
+ "full_attention",
57
+ "full_attention",
58
+ "full_attention",
59
+ "full_attention",
60
+ "full_attention",
61
+ "full_attention",
62
+ "full_attention",
63
+ "full_attention",
64
+ "full_attention",
65
+ "full_attention",
66
+ "full_attention",
67
+ "full_attention",
68
+ "full_attention",
69
+ "full_attention",
70
+ "full_attention",
71
+ "full_attention",
72
+ "full_attention"
73
+ ],
74
+ "max_position_embeddings": 128000,
75
+ "max_window_layers": 28,
76
+ "model_type": "qwen2_5_vl_text",
77
+ "num_attention_heads": 28,
78
+ "num_hidden_layers": 28,
79
+ "num_key_value_heads": 4,
80
+ "pad_token_id": 151643,
81
+ "rms_norm_eps": 1e-06,
82
+ "rope_scaling": {
83
+ "mrope_section": [
84
+ 16,
85
+ 24,
86
+ 24
87
+ ],
88
+ "rope_type": "default",
89
+ "type": "default"
90
+ },
91
+ "rope_theta": 1000000.0,
92
+ "sliding_window": null,
93
+ "use_cache": true,
94
+ "use_sliding_window": false,
95
+ "vision_token_id": 151654,
96
+ "vocab_size": 152064
97
+ },
98
+ "tie_word_embeddings": false,
99
+ "transformers_version": "4.57.1",
100
+ "use_cache": true,
101
+ "use_sliding_window": false,
102
+ "video_token_id": 151656,
103
+ "vision_config": {
104
+ "depth": 32,
105
+ "dtype": "float32",
106
+ "fullatt_block_indexes": [
107
+ 7,
108
+ 15,
109
+ 23,
110
+ 31
111
+ ],
112
+ "hidden_act": "silu",
113
+ "hidden_size": 1280,
114
+ "in_channels": 3,
115
+ "in_chans": 3,
116
+ "initializer_range": 0.02,
117
+ "intermediate_size": 3420,
118
+ "model_type": "qwen2_5_vl",
119
+ "num_heads": 16,
120
+ "out_hidden_size": 3584,
121
+ "patch_size": 14,
122
+ "spatial_merge_size": 2,
123
+ "spatial_patch_size": 14,
124
+ "temporal_patch_size": 2,
125
+ "tokens_per_second": 2,
126
+ "window_size": 112
127
+ },
128
+ "vision_end_token_id": 151653,
129
+ "vision_start_token_id": 151652,
130
+ "vision_token_id": 151654,
131
+ "vocab_size": 152064
132
+ }