gary2oos commited on
Commit
cc718d6
·
verified ·
1 Parent(s): 53ce8bf

Upload 9 files

Browse files
alive_fine-tuning.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452818abbd354bbfe09a35ea8341fd456bf074ad62183897f789ecbadd962af5
3
+ size 724660371
duel_fine-tuning.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ba9fd824992be31218e6214b324081069510a122312c9e57a54d5cec848d99
3
+ size 746661171
nxt_kill_fine-tuning.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a52578fc67b33dbdf1ee266778da8aba52acafc088281d09ef7fd0e27c880a9d
3
+ size 724734291
tfm_alive_fine-tuning.yaml ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model configuration
2
+ model:
3
+ model_name: "TickTransformerModelROPE"
4
+
5
+ vocab_size: 979 # Vocabulary size for token embeddings
6
+ embed_dim: 640 # Embedding dimension
7
+ seq_len: 512 # Sequence length per tick
8
+ dropout: 0.1 # Dropout rate
9
+
10
+ # Embedder (non-causal transformer encoder)
11
+ embedder_heads: 10
12
+ embedder_layers: 6
13
+
14
+ # Processor (GPT-style causal transformer for next token prediction)
15
+ processor_heads: 10
16
+ processor_layers: 8
17
+
18
+ # Decoder (non-causal transformer to decode embeddings to sequences)
19
+ decoder_heads: 10
20
+ decoder_layers: 6
21
+
22
+ # Alive prediction head
23
+ alive_hidden_dim: 512 # Hidden dimension for alive prediction head
24
+ alive_hidden_layers: 1 # Number of hidden layers in alive prediction head
25
+
26
+ # Training configuration
27
+ training:
28
+ batch_size: 16
29
+ grad_accum_steps: 4
30
+ learning_rate_prediction_head: 0.00005
31
+ learning_rate_embedder: 0
32
+ learning_rate_processor: 0.00001
33
+ weight_decay: 0.01
34
+ num_epochs: 30
35
+ warmup_steps: 1000
36
+ max_grad_norm: 1.0
37
+
38
+ scheduler: 'cosine' # 'cosine' or 'linear'
39
+
40
+ base_model_path: '/share/guwanjun-local/cs2-demo-analytics/checkpoints_pretraining_v2/final.pth'
41
+ checkpoint_dir: 'checkpoints_alive_fine-tuning_v2'
42
+
43
+ from_scratch: False # Whether to train from scratch or fine-tune from a pre-trained model
44
+
45
+ use_lora: False # Whether to use LoRA for fine-tuning
46
+ lora_r: 8
47
+ lora_alpha: 16
48
+ lora_dropout: 0.1
49
+
50
+ # Data configuration
51
+ data:
52
+ train_data_path:
53
+ - archive_1.pt
54
+ - new_archive_1.pt
55
+ - archive_2.pt
56
+ - new_archive_2.pt
57
+ - archive_3.pt
58
+ - new_archive_3.pt
59
+ - archive_4.pt
60
+ - new_archive_4.pt
61
+ - new_archive_5.pt
62
+ - archive_5.pt
63
+ - archive_6.pt
64
+ - new_archive_6.pt
65
+ - archive_7.pt
66
+ - new_archive_7.pt
67
+ - archive_8.pt
68
+ - new_archive_8.pt
69
+ - archive_9.pt
70
+ - new_archive_9.pt
71
+ - archive_10.pt
72
+ - new_archive_10.pt
73
+ - archive_11.pt
74
+ - new_archive_11.pt
75
+ - archive_12.pt
76
+ - new_archive_12.pt
77
+ - archive_13.pt
78
+ - new_archive_13.pt
79
+ - archive_14.pt
80
+ - new_archive_14.pt
81
+ - archive_15.pt
82
+ - new_archive_15.pt
83
+ - archive_16.pt
84
+ - new_archive_16.pt
85
+ - archive_17.pt
86
+ - new_archive_17.pt
87
+ - archive_18.pt
88
+ - new_archive_18.pt
89
+ - archive_19.pt
90
+ - new_archive_19.pt
91
+ - archive_20.pt
92
+ - new_archive_20.pt
93
+ - archive_21.pt
94
+ - new_archive_21.pt
95
+ - archive_22.pt
96
+ - new_archive_22.pt
97
+ - archive_23.pt
98
+ - new_archive_23.pt
99
+ - archive_24.pt
100
+ - new_archive_24.pt
101
+ - archive_25.pt
102
+ - new_archive_25.pt
103
+ - archive_26.pt
104
+ - new_archive_26.pt
105
+ - archive_27.pt
106
+ - new_archive_27.pt
107
+ - archive_28.pt
108
+ - new_archive_28.pt
109
+ - archive_29.pt
110
+ - new_archive_29.pt
111
+ - archive_30.pt
112
+ - new_archive_30.pt
113
+ - archive_31.pt
114
+ - new_archive_31.pt
115
+ - archive_32.pt
116
+ - new_archive_32.pt
117
+ - archive_33.pt
118
+ - new_archive_33.pt
119
+ - archive_34.pt
120
+ - new_archive_34.pt
121
+ - archive_35.pt
122
+ - new_archive_35.pt
123
+ - archive_36.pt
124
+ - new_archive_36.pt
125
+ - archive_37.pt
126
+ - new_archive_37.pt
127
+ - archive_38.pt
128
+ - new_archive_38.pt
129
+ - archive_39.pt
130
+ - new_archive_39.pt
131
+ - archive_40.pt
132
+ - new_archive_40.pt
133
+ - archive_41.pt
134
+ - new_archive_41.pt
135
+ - archive_42.pt
136
+ - new_archive_42.pt
137
+ - archive_43.pt
138
+ - new_archive_43.pt
139
+ - archive_44.pt
140
+ - new_archive_44.pt
141
+ - archive_45.pt
142
+ - archive_46.pt
143
+ - archive_47.pt
144
+ - archive_48.pt
145
+ - archive_49.pt
146
+ - archive_50.pt
147
+ - archive_51.pt
148
+ - archive_52.pt
149
+ - archive_53.pt
150
+ - archive_54.pt
151
+ - archive_55.pt
152
+ - archive_56.pt
153
+ - archive_57.pt
154
+ - archive_58.pt
155
+ - archive_59.pt
156
+ - archive_60.pt
157
+ - archive_61.pt
158
+ - archive_106.pt
159
+ - archive_107.pt
160
+ val_data_path:
161
+ - archive_108.pt
162
+ - archive_109.pt
163
+ num_workers: 4
164
+
165
+ # Data dimensions (must match model)
166
+ ticks_per_sample: 64 # Number of ticks in each training sample
167
+ seq_len: 512 # Must match model.seq_len
168
+ pad_token: 978 # Token ID used for padding sequences
169
+
170
+ # Device configuration
171
+ device: 'cuda:1' # 'cuda' or 'cpu'
172
+
173
+ # wandb logging configuration
174
+ logging:
175
+ project_name: 'tick-transformer-alive-fine-tuning'
176
+ test: 2048
177
+
tfm_duel_fine-tuning.yaml ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model configuration
2
+ model:
3
+ model_name: "TickTransformerModelROPE"
4
+
5
+ vocab_size: 979 # Vocabulary size for token embeddings
6
+ embed_dim: 640 # Embedding dimension
7
+ seq_len: 512 # Sequence length per tick
8
+ dropout: 0.1 # Dropout rate
9
+
10
+ # Embedder (non-causal transformer encoder)
11
+ embedder_heads: 10
12
+ embedder_layers: 6
13
+
14
+ # Processor (GPT-style causal transformer for next token prediction)
15
+ processor_heads: 10
16
+ processor_layers: 8
17
+
18
+ # Decoder (non-causal transformer to decode embeddings to sequences)
19
+ decoder_heads: 10
20
+ decoder_layers: 6
21
+
22
+ # Duel prediction head
23
+ duel_hidden_dim: 1024 # Hidden dimension for duel prediction head
24
+ duel_hidden_layers: 2 # Number of hidden layers in duel prediction head
25
+ duel_player_embedding_dim: 64
26
+
27
+ # Training configuration
28
+ training:
29
+ batch_size: 16
30
+ grad_accum_steps: 4
31
+ learning_rate_prediction_head: 0.00005
32
+ learning_rate_embedder: 0
33
+ learning_rate_processor: 0.00001
34
+ weight_decay: 0.01
35
+ num_epochs: 30
36
+ warmup_steps: 1000
37
+ max_grad_norm: 1.0
38
+
39
+ scheduler: 'cosine' # 'cosine' or 'linear'
40
+
41
+ base_model_path: '/share/guwanjun-local/cs2-demo-analytics/checkpoints_pretraining_v2/final.pth'
42
+ checkpoint_dir: 'checkpoints_duel_fine-tuning_v2'
43
+
44
+ from_scratch: False # Whether to train from scratch or fine-tune from a pre-trained model
45
+
46
+ use_lora: False # Whether to use LoRA for fine-tuning
47
+ lora_r: 8
48
+ lora_alpha: 16
49
+ lora_dropout: 0.1
50
+
51
+ # Data configuration
52
+ data:
53
+ train_data_path:
54
+ - archive_1.pt
55
+ - new_archive_1.pt
56
+ - archive_2.pt
57
+ - new_archive_2.pt
58
+ - archive_3.pt
59
+ - new_archive_3.pt
60
+ - archive_4.pt
61
+ - new_archive_4.pt
62
+ - new_archive_5.pt
63
+ - archive_5.pt
64
+ - archive_6.pt
65
+ - new_archive_6.pt
66
+ - archive_7.pt
67
+ - new_archive_7.pt
68
+ - archive_8.pt
69
+ - new_archive_8.pt
70
+ - archive_9.pt
71
+ - new_archive_9.pt
72
+ - archive_10.pt
73
+ - new_archive_10.pt
74
+ - archive_11.pt
75
+ - new_archive_11.pt
76
+ - archive_12.pt
77
+ - new_archive_12.pt
78
+ - archive_13.pt
79
+ - new_archive_13.pt
80
+ - archive_14.pt
81
+ - new_archive_14.pt
82
+ - archive_15.pt
83
+ - new_archive_15.pt
84
+ - archive_16.pt
85
+ - new_archive_16.pt
86
+ - archive_17.pt
87
+ - new_archive_17.pt
88
+ - archive_18.pt
89
+ - new_archive_18.pt
90
+ - archive_19.pt
91
+ - new_archive_19.pt
92
+ - archive_20.pt
93
+ - new_archive_20.pt
94
+ - archive_21.pt
95
+ - new_archive_21.pt
96
+ - archive_22.pt
97
+ - new_archive_22.pt
98
+ - archive_23.pt
99
+ - new_archive_23.pt
100
+ - archive_24.pt
101
+ - new_archive_24.pt
102
+ - archive_25.pt
103
+ - new_archive_25.pt
104
+ - archive_26.pt
105
+ - new_archive_26.pt
106
+ - archive_27.pt
107
+ - new_archive_27.pt
108
+ - archive_28.pt
109
+ - new_archive_28.pt
110
+ - archive_29.pt
111
+ - new_archive_29.pt
112
+ - archive_30.pt
113
+ - new_archive_30.pt
114
+ - archive_31.pt
115
+ - new_archive_31.pt
116
+ - archive_32.pt
117
+ - new_archive_32.pt
118
+ - archive_33.pt
119
+ - new_archive_33.pt
120
+ - archive_34.pt
121
+ - new_archive_34.pt
122
+ - archive_35.pt
123
+ - new_archive_35.pt
124
+ - archive_36.pt
125
+ - new_archive_36.pt
126
+ - archive_37.pt
127
+ - new_archive_37.pt
128
+ - archive_38.pt
129
+ - new_archive_38.pt
130
+ - archive_39.pt
131
+ - new_archive_39.pt
132
+ - archive_40.pt
133
+ - new_archive_40.pt
134
+ - archive_41.pt
135
+ - new_archive_41.pt
136
+ - archive_42.pt
137
+ - new_archive_42.pt
138
+ - archive_43.pt
139
+ - new_archive_43.pt
140
+ - archive_44.pt
141
+ - new_archive_44.pt
142
+ - archive_45.pt
143
+ - archive_46.pt
144
+ - archive_47.pt
145
+ - archive_48.pt
146
+ - archive_49.pt
147
+ - archive_50.pt
148
+ - archive_51.pt
149
+ - archive_52.pt
150
+ - archive_53.pt
151
+ - archive_54.pt
152
+ - archive_55.pt
153
+ - archive_56.pt
154
+ - archive_57.pt
155
+ - archive_58.pt
156
+ - archive_59.pt
157
+ - archive_60.pt
158
+ - archive_61.pt
159
+ - archive_106.pt
160
+ - archive_107.pt
161
+ val_data_path:
162
+ - archive_108.pt
163
+ - archive_109.pt
164
+ num_workers: 4
165
+
166
+ # Data dimensions (must match model)
167
+ ticks_per_sample: 64 # Number of ticks in each training sample
168
+ seq_len: 512 # Must match model.seq_len
169
+ pad_token: 978 # Token ID used for padding sequences
170
+
171
+ # Device configuration
172
+ device: 'cuda:3' # 'cuda' or 'cpu'
173
+
174
+ # wandb logging configuration
175
+ logging:
176
+ project_name: 'tick-transformer-duel-fine-tuning'
177
+ test: 2048
178
+
tfm_nxt_kill_fine-tuning.yaml ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model configuration
2
+ model:
3
+ model_name: "TickTransformerModelROPE"
4
+
5
+ vocab_size: 979 # Vocabulary size for token embeddings
6
+ embed_dim: 640 # Embedding dimension
7
+ seq_len: 512 # Sequence length per tick
8
+ dropout: 0.1 # Dropout rate
9
+
10
+ # Embedder (non-causal transformer encoder)
11
+ embedder_heads: 10
12
+ embedder_layers: 6
13
+
14
+ # Processor (GPT-style causal transformer for next token prediction)
15
+ processor_heads: 10
16
+ processor_layers: 8
17
+
18
+ # Decoder (non-causal transformer to decode embeddings to sequences)
19
+ decoder_heads: 10
20
+ decoder_layers: 6
21
+
22
+ # nxt_kill prediction head
23
+ nxt_kill_hidden_dim: 512 # Hidden dimension for nxt_kill prediction head
24
+ nxt_kill_hidden_layers: 1 # Number of hidden layers in nxt_kill prediction head
25
+
26
+ # Training configuration
27
+ training:
28
+ batch_size: 16
29
+ grad_accum_steps: 4
30
+ learning_rate_prediction_head: 0.00005
31
+ learning_rate_embedder: 0
32
+ learning_rate_processor: 0.00001
33
+ weight_decay: 0.01
34
+ num_epochs: 30
35
+ warmup_steps: 1000
36
+ max_grad_norm: 1.0
37
+
38
+ scheduler: 'cosine' # 'cosine' or 'linear'
39
+
40
+ base_model_path: '/share/guwanjun-local/cs2-demo-analytics/checkpoints_pretraining_v2/final.pth'
41
+ checkpoint_dir: 'checkpoints_nxt_kill_fine-tuning_v2'
42
+
43
+ from_scratch: False # Whether to train from scratch or fine-tune from a pre-trained model
44
+
45
+ use_lora: False # Whether to use LoRA for fine-tuning
46
+ lora_r: 8
47
+ lora_alpha: 16
48
+ lora_dropout: 0.1
49
+
50
+ # Data configuration
51
+ data:
52
+ train_data_path:
53
+ - archive_1.pt
54
+ - new_archive_1.pt
55
+ - archive_2.pt
56
+ - new_archive_2.pt
57
+ - archive_3.pt
58
+ - new_archive_3.pt
59
+ - archive_4.pt
60
+ - new_archive_4.pt
61
+ - new_archive_5.pt
62
+ - archive_5.pt
63
+ - archive_6.pt
64
+ - new_archive_6.pt
65
+ - archive_7.pt
66
+ - new_archive_7.pt
67
+ - archive_8.pt
68
+ - new_archive_8.pt
69
+ - archive_9.pt
70
+ - new_archive_9.pt
71
+ - archive_10.pt
72
+ - new_archive_10.pt
73
+ - archive_11.pt
74
+ - new_archive_11.pt
75
+ - archive_12.pt
76
+ - new_archive_12.pt
77
+ - archive_13.pt
78
+ - new_archive_13.pt
79
+ - archive_14.pt
80
+ - new_archive_14.pt
81
+ - archive_15.pt
82
+ - new_archive_15.pt
83
+ - archive_16.pt
84
+ - new_archive_16.pt
85
+ - archive_17.pt
86
+ - new_archive_17.pt
87
+ - archive_18.pt
88
+ - new_archive_18.pt
89
+ - archive_19.pt
90
+ - new_archive_19.pt
91
+ - archive_20.pt
92
+ - new_archive_20.pt
93
+ - archive_21.pt
94
+ - new_archive_21.pt
95
+ - archive_22.pt
96
+ - new_archive_22.pt
97
+ - archive_23.pt
98
+ - new_archive_23.pt
99
+ - archive_24.pt
100
+ - new_archive_24.pt
101
+ - archive_25.pt
102
+ - new_archive_25.pt
103
+ - archive_26.pt
104
+ - new_archive_26.pt
105
+ - archive_27.pt
106
+ - new_archive_27.pt
107
+ - archive_28.pt
108
+ - new_archive_28.pt
109
+ - archive_29.pt
110
+ - new_archive_29.pt
111
+ - archive_30.pt
112
+ - new_archive_30.pt
113
+ - archive_31.pt
114
+ - new_archive_31.pt
115
+ - archive_32.pt
116
+ - new_archive_32.pt
117
+ - archive_33.pt
118
+ - new_archive_33.pt
119
+ - archive_34.pt
120
+ - new_archive_34.pt
121
+ - archive_35.pt
122
+ - new_archive_35.pt
123
+ - archive_36.pt
124
+ - new_archive_36.pt
125
+ - archive_37.pt
126
+ - new_archive_37.pt
127
+ - archive_38.pt
128
+ - new_archive_38.pt
129
+ - archive_39.pt
130
+ - new_archive_39.pt
131
+ - archive_40.pt
132
+ - new_archive_40.pt
133
+ - archive_41.pt
134
+ - new_archive_41.pt
135
+ - archive_42.pt
136
+ - new_archive_42.pt
137
+ - archive_43.pt
138
+ - new_archive_43.pt
139
+ - archive_44.pt
140
+ - new_archive_44.pt
141
+ - archive_45.pt
142
+ - archive_46.pt
143
+ - archive_47.pt
144
+ - archive_48.pt
145
+ - archive_49.pt
146
+ - archive_50.pt
147
+ - archive_51.pt
148
+ - archive_52.pt
149
+ - archive_53.pt
150
+ - archive_54.pt
151
+ - archive_55.pt
152
+ - archive_56.pt
153
+ - archive_57.pt
154
+ - archive_58.pt
155
+ - archive_59.pt
156
+ - archive_60.pt
157
+ - archive_61.pt
158
+ - archive_106.pt
159
+ - archive_107.pt
160
+ val_data_path:
161
+ - archive_108.pt
162
+ - archive_109.pt
163
+ num_workers: 4
164
+
165
+ # Data dimensions (must match model)
166
+ ticks_per_sample: 64 # Number of ticks in each training sample
167
+ seq_len: 512 # Must match model.seq_len
168
+ pad_token: 978 # Token ID used for padding sequences
169
+
170
+ # Device configuration
171
+ device: 'cuda:2' # 'cuda' or 'cpu'
172
+
173
+ # wandb logging configuration
174
+ logging:
175
+ project_name: 'tick-transformer-nxt_kill-fine-tuning'
176
+ test: 2048
177
+
tfm_win_rate_fine-tuning.yaml ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model configuration
2
+ model:
3
+ model_name: "TickTransformerModelROPE"
4
+
5
+ vocab_size: 979 # Vocabulary size for token embeddings
6
+ embed_dim: 640 # Embedding dimension
7
+ seq_len: 512 # Sequence length per tick
8
+ dropout: 0.1 # Dropout rate
9
+
10
+ # Embedder (non-causal transformer encoder)
11
+ embedder_heads: 10
12
+ embedder_layers: 6
13
+
14
+ # Processor (GPT-style causal transformer for next token prediction)
15
+ processor_heads: 10
16
+ processor_layers: 8
17
+
18
+ # Decoder (non-causal transformer to decode embeddings to sequences)
19
+ decoder_heads: 10
20
+ decoder_layers: 6
21
+
22
+ # Win rate prediction head
23
+ win_rate_hidden_dim: 512 # Hidden dimension for win rate prediction head
24
+ win_rate_hidden_layers: 1 # Number of hidden layers in win rate prediction head
25
+
26
+ # Training configuration
27
+ training:
28
+ batch_size: 16
29
+ grad_accum_steps: 4
30
+ learning_rate_prediction_head: 0.00005
31
+ learning_rate_embedder: 0
32
+ learning_rate_processor: 0.00001
33
+ weight_decay: 0.01
34
+ num_epochs: 30
35
+ warmup_steps: 1000
36
+ max_grad_norm: 1.0
37
+
38
+ scheduler: 'cosine' # 'cosine' or 'linear'
39
+
40
+ base_model_path: '/share/guwanjun-local/cs2-demo-analytics/checkpoints_pretraining_v2/final.pth'
41
+ checkpoint_dir: 'checkpoints_win_rate_fine-tuning_v2'
42
+
43
+ from_scratch: False # Whether to train from scratch or fine-tune from a pre-trained model
44
+
45
+ use_lora: False # Whether to use LoRA for fine-tuning
46
+ lora_r: 8
47
+ lora_alpha: 16
48
+ lora_dropout: 0.1
49
+
50
+ # Data configuration
51
+ data:
52
+ train_data_path:
53
+ - archive_1.pt
54
+ - new_archive_1.pt
55
+ - archive_2.pt
56
+ - new_archive_2.pt
57
+ - archive_3.pt
58
+ - new_archive_3.pt
59
+ - archive_4.pt
60
+ - new_archive_4.pt
61
+ - new_archive_5.pt
62
+ - archive_5.pt
63
+ - archive_6.pt
64
+ - new_archive_6.pt
65
+ - archive_7.pt
66
+ - new_archive_7.pt
67
+ - archive_8.pt
68
+ - new_archive_8.pt
69
+ - archive_9.pt
70
+ - new_archive_9.pt
71
+ - archive_10.pt
72
+ - new_archive_10.pt
73
+ - archive_11.pt
74
+ - new_archive_11.pt
75
+ - archive_12.pt
76
+ - new_archive_12.pt
77
+ - archive_13.pt
78
+ - new_archive_13.pt
79
+ - archive_14.pt
80
+ - new_archive_14.pt
81
+ - archive_15.pt
82
+ - new_archive_15.pt
83
+ - archive_16.pt
84
+ - new_archive_16.pt
85
+ - archive_17.pt
86
+ - new_archive_17.pt
87
+ - archive_18.pt
88
+ - new_archive_18.pt
89
+ - archive_19.pt
90
+ - new_archive_19.pt
91
+ - archive_20.pt
92
+ - new_archive_20.pt
93
+ - archive_21.pt
94
+ - new_archive_21.pt
95
+ - archive_22.pt
96
+ - new_archive_22.pt
97
+ - archive_23.pt
98
+ - new_archive_23.pt
99
+ - archive_24.pt
100
+ - new_archive_24.pt
101
+ - archive_25.pt
102
+ - new_archive_25.pt
103
+ - archive_26.pt
104
+ - new_archive_26.pt
105
+ - archive_27.pt
106
+ - new_archive_27.pt
107
+ - archive_28.pt
108
+ - new_archive_28.pt
109
+ - archive_29.pt
110
+ - new_archive_29.pt
111
+ - archive_30.pt
112
+ - new_archive_30.pt
113
+ - archive_31.pt
114
+ - new_archive_31.pt
115
+ - archive_32.pt
116
+ - new_archive_32.pt
117
+ - archive_33.pt
118
+ - new_archive_33.pt
119
+ - archive_34.pt
120
+ - new_archive_34.pt
121
+ - archive_35.pt
122
+ - new_archive_35.pt
123
+ - archive_36.pt
124
+ - new_archive_36.pt
125
+ - archive_37.pt
126
+ - new_archive_37.pt
127
+ - archive_38.pt
128
+ - new_archive_38.pt
129
+ - archive_39.pt
130
+ - new_archive_39.pt
131
+ - archive_40.pt
132
+ - new_archive_40.pt
133
+ - archive_41.pt
134
+ - new_archive_41.pt
135
+ - archive_42.pt
136
+ - new_archive_42.pt
137
+ - archive_43.pt
138
+ - new_archive_43.pt
139
+ - archive_44.pt
140
+ - new_archive_44.pt
141
+ - archive_45.pt
142
+ - archive_46.pt
143
+ - archive_47.pt
144
+ - archive_48.pt
145
+ - archive_49.pt
146
+ - archive_50.pt
147
+ - archive_51.pt
148
+ - archive_52.pt
149
+ - archive_53.pt
150
+ - archive_54.pt
151
+ - archive_55.pt
152
+ - archive_56.pt
153
+ - archive_57.pt
154
+ - archive_58.pt
155
+ - archive_59.pt
156
+ - archive_60.pt
157
+ - archive_61.pt
158
+ - archive_106.pt
159
+ - archive_107.pt
160
+ val_data_path:
161
+ - archive_108.pt
162
+ - archive_109.pt
163
+ num_workers: 4
164
+
165
+ # Data dimensions (must match model)
166
+ ticks_per_sample: 64 # Number of ticks in each training sample
167
+ seq_len: 512 # Must match model.seq_len
168
+ pad_token: 978 # Token ID used for padding sequences
169
+
170
+ # Device configuration
171
+ device: 'cuda:0' # 'cuda' or 'cpu'
172
+
173
+ # wandb logging configuration
174
+ logging:
175
+ project_name: 'tick-transformer-win-rate-fine-tuning'
176
+ test: 2048
177
+
tokenizer.yaml ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ maps:
2
+ de_mirage:
3
+ center:
4
+ - -605.8900146484375
5
+ - -866.8900146484375
6
+ - -171.6199951171875
7
+ de_dust2:
8
+ center:
9
+ - -199.0
10
+ - 977.0
11
+ - 32.220001220703125
12
+ de_inferno:
13
+ center:
14
+ - 481.07000732421875
15
+ - 1396.47998046875
16
+ - 137.91000366210938
17
+ de_nuke:
18
+ center:
19
+ - 265.9599914550781
20
+ - -772.5
21
+ - -381.8999938964844
22
+ de_overpass:
23
+ center:
24
+ - -2027.3900146484375
25
+ - -812.9000244140625
26
+ - 324.95001220703125
27
+ de_ancient:
28
+ center:
29
+ - -435.5
30
+ - -348.0
31
+ - 43.650001525878906
32
+ de_anubis:
33
+ center:
34
+ - -77.38999938964844
35
+ - 618.9000244140625
36
+ - -6.800000190734863
37
+ de_train:
38
+ center:
39
+ - -118.25
40
+ - -2.0
41
+ - -128.52000427246094
42
+
43
+ weapons:
44
+ - Desert Eagle
45
+ - Dual Berettas
46
+ - Five-SeveN
47
+ - Glock-18
48
+ - AK-47
49
+ - AUG
50
+ - AWP
51
+ - FAMAS
52
+ - G3SG1
53
+ - Galil AR
54
+ - M249
55
+ - M4A4
56
+ - MAC-10
57
+ - P90
58
+ - MP5-SD
59
+ - UMP-45
60
+ - XM1014
61
+ - PP-Bizon
62
+ - MAG-7
63
+ - Negev
64
+ - Sawed-Off
65
+ - Tec-9
66
+ - Zeus x27
67
+ - P2000
68
+ - MP7
69
+ - MP9
70
+ - Nova
71
+ - P250
72
+ - SCAR-20
73
+ - SG 553
74
+ - SSG 08
75
+ - knife
76
+ - Flashbang
77
+ - High Explosive Grenade
78
+ - Smoke Grenade
79
+ - Molotov
80
+ - Decoy Grenade
81
+ - Incendiary Grenade
82
+ - C4 Explosive
83
+ - Kevlar Vest
84
+ - Kevlar & Helmet
85
+ - Heavy Assault Suit
86
+ - item_nvg
87
+ - Defuse Kit
88
+ - Rescue Kit
89
+ - Medi-Shot
90
+ - M4A1-S
91
+ - USP-S
92
+ - Trade Up Contract
93
+ - CZ75-Auto
94
+ - R8 Revolver
95
+
96
+ projectiles: # 落地后的状态
97
+ - smokegrenade
98
+ - inferno
99
+
100
+ entity_projectiles: # 飞行中的状态
101
+ - CHEGrenadeProjectile
102
+ - CSmokeGrenadeProjectile
103
+ - CMolotovProjectile
104
+ - CDecoyProjectile
105
+ - CFlashbangProjectile
106
+
107
+ safe_mode: 0
108
+
109
+ tokenizer:
110
+ position:
111
+ x:
112
+ range: [-5000, 5000]
113
+ grid: 1 # tokenizer离散化每个单位 = 1 个游戏单位
114
+ y:
115
+ range: [-5000, 5000]
116
+ grid: 1
117
+ z:
118
+ range: [-2000, 2000]
119
+ grid: 1
120
+
121
+ view_angle:
122
+ pitch:
123
+ range: [-90, 90]
124
+ grid: 1
125
+ yaw:
126
+ range: [-180, 180]
127
+ grid: 1
128
+
129
+ status:
130
+ health_grid: 1
131
+ time_grid: 1
132
+ max_time: 170
133
+ max_health: 100
134
+ n_players: 10
135
+
136
+ planted_duration:
137
+ range: [0, 40]
138
+ grid: 0.5
win_rate_fine-tuning.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:612598103ef8b8b96d60add6ab968ec7ac0123b68372ac37713de5879d3c57d0
3
+ size 724605075