OpenOneRec Technical Report
Paper • 2512.24762 • Published
Qwen3-1.8B fully fine-tuned for generative product recommendation via Semantic IDs. Trained on new pipeline data (63K items, 4.7M conversations). This is the baseline model (0% general data mixing) for the H2 experiment comparing the effect of reasoning data mixing.
Epoch 0.0: loss 2.956
Epoch 0.5: loss 1.80
Epoch 1.0: loss 1.68
Epoch 2.0: loss 1.63
Epoch 3.0: loss 1.59 (final: 1.687 avg)
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("kalistratov/qwen3-1.8b-sid-baseline")
tokenizer = AutoTokenizer.from_pretrained("kalistratov/qwen3-1.8b-sid-baseline")
This model is part of hypothesis H2: "Does reasoning data mixing improve SID prediction quality?"
| Variant | General data | Status |
|---|---|---|
| This model (baseline) | 0% | Completed |
| + 25% reasoning | OpenMathReasoning + OpenCodeReasoning + reasoning-v1 | Pending |
Master's thesis, Moscow Institute of Physics and Technology (MIPT), 2026.