ritishshrirao commited on
Commit
3e893cd
·
1 Parent(s): e44cdee

Minor config changes

Browse files
config/self_play_training_hf_a10g_smoke.json CHANGED
@@ -41,8 +41,8 @@
41
  "loss_type": "dapo",
42
  "scale_rewards": "group",
43
  "logging_steps": 1,
44
- "save_steps": 10,
45
- "save_total_limit": 2,
46
  "optim": "adamw_torch_fused",
47
  "bf16": true,
48
  "tf32": true,
@@ -73,8 +73,8 @@
73
  "loss_type": "dapo",
74
  "scale_rewards": "group",
75
  "logging_steps": 1,
76
- "save_steps": 10,
77
- "save_total_limit": 2,
78
  "optim": "adamw_torch_fused",
79
  "bf16": true,
80
  "tf32": true,
 
41
  "loss_type": "dapo",
42
  "scale_rewards": "group",
43
  "logging_steps": 1,
44
+ "save_steps": 50,
45
+ "save_total_limit": 1,
46
  "optim": "adamw_torch_fused",
47
  "bf16": true,
48
  "tf32": true,
 
73
  "loss_type": "dapo",
74
  "scale_rewards": "group",
75
  "logging_steps": 1,
76
+ "save_steps": 50,
77
+ "save_total_limit": 1,
78
  "optim": "adamw_torch_fused",
79
  "bf16": true,
80
  "tf32": true,
config/self_play_training_hf_l40s_full.json CHANGED
@@ -41,8 +41,8 @@
41
  "loss_type": "dapo",
42
  "scale_rewards": "group",
43
  "logging_steps": 5,
44
- "save_steps": 30,
45
- "save_total_limit": 4,
46
  "optim": "adamw_torch_fused",
47
  "bf16": true,
48
  "tf32": true,
@@ -73,8 +73,8 @@
73
  "loss_type": "dapo",
74
  "scale_rewards": "group",
75
  "logging_steps": 5,
76
- "save_steps": 30,
77
- "save_total_limit": 4,
78
  "optim": "adamw_torch_fused",
79
  "bf16": true,
80
  "tf32": true,
 
41
  "loss_type": "dapo",
42
  "scale_rewards": "group",
43
  "logging_steps": 5,
44
+ "save_steps": 120,
45
+ "save_total_limit": 1,
46
  "optim": "adamw_torch_fused",
47
  "bf16": true,
48
  "tf32": true,
 
73
  "loss_type": "dapo",
74
  "scale_rewards": "group",
75
  "logging_steps": 5,
76
+ "save_steps": 120,
77
+ "save_total_limit": 1,
78
  "optim": "adamw_torch_fused",
79
  "bf16": true,
80
  "tf32": true,
src/osint_env/training/self_play.py CHANGED
@@ -114,13 +114,26 @@ def _maybe_upload_folder_to_hf(local_dir: Path, run_dir: Path, commit_message: s
114
  path_in_repo = _hf_relative_repo_path(local_dir, run_dir)
115
  api = HfApi(token=token)
116
  api.create_repo(repo_id=repo_id, repo_type=repo_type, private=private, exist_ok=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
117
  api.upload_folder(
118
  folder_path=str(local_dir),
119
  repo_id=repo_id,
120
  repo_type=repo_type,
121
  path_in_repo=path_in_repo,
122
  commit_message=commit_message,
123
- ignore_patterns=["*.pyc", "__pycache__", ".DS_Store"],
124
  )
125
  print(f"[self_play][hf_upload] uploaded {local_dir} -> {repo_type}:{repo_id}/{path_in_repo}")
126
 
 
114
  path_in_repo = _hf_relative_repo_path(local_dir, run_dir)
115
  api = HfApi(token=token)
116
  api.create_repo(repo_id=repo_id, repo_type=repo_type, private=private, exist_ok=True)
117
+ # Upload only inference-relevant artifacts. Resume-only state such as
118
+ # optimizer/scheduler RNG snapshots makes uploads much larger and is not
119
+ # needed for sharing or post-phase evaluation.
120
+ ignore_patterns = [
121
+ "*.pyc",
122
+ "__pycache__",
123
+ ".DS_Store",
124
+ "**/optimizer.pt",
125
+ "**/scheduler.pt",
126
+ "**/rng_state.pth",
127
+ "**/trainer_state.json",
128
+ "**/training_args.bin",
129
+ ]
130
  api.upload_folder(
131
  folder_path=str(local_dir),
132
  repo_id=repo_id,
133
  repo_type=repo_type,
134
  path_in_repo=path_in_repo,
135
  commit_message=commit_message,
136
+ ignore_patterns=ignore_patterns,
137
  )
138
  print(f"[self_play][hf_upload] uploaded {local_dir} -> {repo_type}:{repo_id}/{path_in_repo}")
139