Update README.md
ff7e304 verified - 1.52 kB initial commit
- 4.96 kB Update README.md
- 440 Bytes Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 718 Bytes Upload LlamaForCausalLM
- 224 Bytes Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 121 Bytes Upload LlamaForCausalLM
- 4.99 GB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 4.98 GB Upload LlamaForCausalLM
- 4.98 GB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 4.99 GB Upload LlamaForCausalLM
- 3.85 GB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 4.92 GB Upload LlamaForCausalLM
- 4.86 GB Upload LlamaForCausalLM
- 4.86 GB Upload LlamaForCausalLM
- 3.03 GB Upload LlamaForCausalLM
- 22.5 kB Upload LlamaForCausalLM
- 369 Bytes Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 4.61 MB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 1.23 kB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 236 Bytes Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
- 733 kB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint
training_args.bin Detected Pickle imports (13)
- "transformers.training_args.OptimizerNames",
- "transformers.integrations.deepspeed.HfTrainerDeepSpeedConfig",
- "transformers.trainer_utils.IntervalStrategy",
- "accelerate.utils.dataclasses.DistributedType",
- "transformers.trainer_utils.SchedulerType",
- "torch.device",
- "accelerate.state.PartialState",
- "transformers.trainer_utils.HubStrategy",
- "torch.bfloat16",
- "h4.training.configs.sft_config.SFTConfig",
- "accelerate.utils.dataclasses.DeepSpeedPlugin",
- "transformers.integrations.deepspeed.HfDeepSpeedConfig",
- "transformers.trainer_pt_utils.AcceleratorConfig"
How to fix it?
6.9 kB Add AI-MO/deepseek-math-7b-sft-aimo_v31.24 checkpoint