Smollm3-3B aligned using DPO on the argilla/ultrafeedback-binarized-preferences
- Downloads last month
- 623
Model tree for MInAlA/SmolLM3-3B-DPO-merged
Base model
HuggingFaceTB/SmolLM3-3B-Base Finetuned
HuggingFaceTB/SmolLM3-3BSmollm3-3B aligned using DPO on the argilla/ultrafeedback-binarized-preferences
Base model
HuggingFaceTB/SmolLM3-3B-Base