For this one...

... (over)trained a SmolLM2-360M on 5 epochs at swept-for LR and rank on each of the target domains to fit style, then rewarded the model for lowering perplexity on the proxy model.

In this case, trained an adapter per domain and then Karcher merged them. I'm not sure if any of the domains had notably different effect, they all basically had the same result on evals. However, the karcher combination of them seem to have significantly lowered perplexity on lambada_openai, which is interesting enough to publish.

Additionally, attempted to implement MARA from https://im-ant.github.io/mara/ on the GRPO side to help preserve distribution entropy, though I'm unsure how correctly/usefully we did so.

Task Metric Qwen3-4B-Base GRPO-Merge Δ Base GRPO-Wave Δ Base Δ Merge Style-Karcher Δ Base Δ Wave
arc_easy acc 0.7891 0.7870 -0.27% 0.7912 +0.27% +0.53% 0.7883 -0.10% -0.37%
arc_easy acc_norm 0.7609 0.7605 -0.05% 0.7643 +0.45% +0.50% 0.7576 -0.43% -1.04%
lambada_openai acc 0.6912 0.6984 +1.04% 0.7006 +1.36% +0.31% 0.7087 +2.53% +1.16%
lambada_openai perplexity ↓ 4.2433 4.0490 -4.58% 3.9616 -6.64% -2.16% 3.8343 -9.63% -3.21%
openbookqa acc 0.3160 0.3180 +0.63% 0.3180 +0.63% ±0.00% 0.3160 ±0.00% -0.63%
openbookqa acc_norm 0.4100 0.4120 +0.49% 0.4100 ±0.00% -0.49% 0.4080 -0.49% -0.49%
piqa acc 0.7797 0.7807 +0.13% 0.7813 +0.21% +0.08% 0.7786 -0.14% -0.35%
piqa acc_norm 0.7807 0.7807 ±0.00% 0.7813 +0.08% +0.08% 0.7807 ±0.00% -0.08%

Some very interesting results on diversity also:

Diversity Metrics (Qwen3-4B-Base vs Style-Karcher, temperature=1.0, 8 completions per prompt)

Domain Metric Base Karcher Δ
ao3_english Prefix entropy 3.309 3.238 -2.1%
ao3_english Distinct-1 0.618 0.683 +10.5%
ao3_english Distinct-2 0.962 0.984 +2.3%
ao3_english Pairwise diversity 0.919 0.932 +1.4%
github_python Prefix entropy 1.514 1.456 -3.8%
github_python Distinct-1 0.610 0.624 +2.3%
github_python Distinct-2 0.890 0.876 -1.6%
github_python Pairwise diversity 0.933 0.933 ±0.0%
wikipedia_english Prefix entropy 1.974 1.892 -4.2%
wikipedia_english Distinct-1 0.599 0.559 -6.7%
wikipedia_english Distinct-2 0.932 0.898 -3.6%
wikipedia_english Pairwise diversity 0.907 0.900 -0.8%
bbc_news Prefix entropy 2.252 2.186 -2.9%
bbc_news Distinct-1 0.557 0.577 +3.6%
bbc_news Distinct-2 0.949 0.951 +0.3%
bbc_news Pairwise diversity 0.901 0.908 +0.8%
arxiv_cs Prefix entropy 2.455 2.346 -4.4%
arxiv_cs Distinct-1 0.555 0.567 +2.3%
arxiv_cs Distinct-2 0.905 0.906 +0.2%
arxiv_cs Pairwise diversity 0.895 0.901 +0.7%

Additional experiment (after quantization, should affect further training but not existing quants): Initializing the <think></think> tokens in embedding space.

Original embeddings were identical (cos=1.0) at 0.3x norm, untrained.

Optimized via AdamW on GSM8k reasoning traces with 3-shot prefix, loss on reasoning+answer tokens, norm clamped to 1.5x avg embedding norm.

After: two distinct vectors (cos=0.07) at 1.5x norm. GSM8k 3-shot accuracy: 96.7% (29/30) vs 90.0% with original embeddings. CE loss improvement: +7.8% on held-out eval.

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Karcher Mean merge method.

Models Merged

The following models were included in the merge:

  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-github_javascript-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-arxiv_cs-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-general-ao3style-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-ao3_english-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-arxiv_math-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-github_python-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-wikipedia_english-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-arxiv_physics-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-github_cpp-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-bbc_news-mara-360m
  • ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave + ../rlvr-envs/grpo-github_markdown-mara-360m

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-ao3_english-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-arxiv_cs-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-arxiv_math-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-arxiv_physics-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-bbc_news-mara-360m

  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-github_cpp-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-github_javascript-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-github_markdown-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-github_python-mara-360m
  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-wikipedia_english-mara-360m

  - model: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave+../rlvr-envs/grpo-general-ao3style-360m
merge_method: karcher
dtype: bfloat16
tokenizer_source: ../rlvr-envs/Qwen3-4B-Base-Continued-GRPO-Wave
Downloads last month
23
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Lambent/Qwen3-4B-Base-Continued-GRPO-Style-Karcher

Finetuned
(1)
this model
Adapters
1 model
Quantizations
2 models