SOKRATES: Qwen3-8B PrOntoQA SFT (Optionized)

Supervised fine-tuned Qwen3-8B for logical reasoning using the optionized Thought/Action format.

Performance

Metric Value
PrOntoQA Accuracy 93.3%

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "Moonlight556/sokrates-qwen3-8b-prontoqa-sft-optionized",
    torch_dtype="bfloat16",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(
    "Moonlight556/sokrates-qwen3-8b-prontoqa-sft-optionized"
)

Related Models

  • DPO Iter 1: sokrates-qwen3-8b-prontoqa-oak-dpo-iter1 (96.8%)
  • DPO Iter 2: sokrates-qwen3-8b-prontoqa-oak-dpo-iter2 (98.1%)
  • DPO Iter 3: sokrates-qwen3-8b-prontoqa-oak-dpo-iter3 (98.2% - best)
Downloads last month
3
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Moonlight556/sokrates-qwen3-8b-prontoqa-sft-optionized

Finetuned
Qwen/Qwen3-8B
Finetuned
(1466)
this model