VERL Math Transfer 7B to 3B fix07 pool7to1

Math transfer experiment trained with verl. This repo groups all exported Hugging Face checkpoints for the 7B-to-3B fix_0_7 pool7to1 configuration.

Layout

  • main: latest exported checkpoint, currently step-070
  • step revisions: step-010, step-020, step-030, step-040, step-050, step-060, step-070

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

repo_id = "hyunseoki/verl-math-transfer-7bi-to-3bi-fix07-pool7to1"
tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(repo_id, trust_remote_code=True)

Load a specific checkpoint revision:

from transformers import AutoTokenizer, AutoModelForCausalLM

repo_id = "hyunseoki/verl-math-transfer-7bi-to-3bi-fix07-pool7to1"
revision = "step-070"
tokenizer = AutoTokenizer.from_pretrained(repo_id, revision=revision, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(repo_id, revision=revision, trust_remote_code=True)

Notes

  • Architecture detected from the exported config: Qwen2ForCausalLM
  • The original base model Hub ID is not encoded in these local checkpoints, so base_model metadata is not set automatically.
  • Checkpoints were exported from verl FSDP shards into Hugging Face safetensors format.
Downloads last month
836
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including hyunseoki/verl-math-transfer-7bi-to-3bi-fix07-pool7to1