C64 Ministral 3 14B Thinking C64 Reasoning
Collection
LoRA and GGUF artifacts for the C64-focused Ministral 3 14B reasoning fine-tune. • 2 items • Updated
This repository contains the LoRA adapter produced by fine-tuning a reasoning-capable Ministral 3 14B model on technical Commodore 64 material.
Objective:
Project source code and pipeline:
Related repositories:
mistralai/Ministral-3-14B-Reasoning-2512Mistral3ForConditionalGeneration (language component fine-tuned with LoRA)text_config.max_position_embeddings)Run summary:
Validation status: FAIL
Source artifacts: results/reasoning_validation/14b/20260302_152057
Note: contract/format retention passed; failure is due to strict exact-token determinism (hash mismatch across repeated same-seed runs).
| Metric | Value |
|---|---|
| single_think_tag_rate | 1.0000 |
| single_balanced_tag_rate | 1.0000 |
| single_final_after_think_rate | 1.0000 |
| multi_turn_retention_rate | 1.0000 |
| format_contract_pass_rate | 1.0000 |
| exact_hash_match_rate | 0.3403 |
| semantic_similarity_avg | 0.9956 |
| crash_or_timeout_rate | 0.0000 |
import torch
from peft import PeftModel
from transformers import AutoTokenizer
from transformers.models.mistral3 import Mistral3ForConditionalGeneration
base_id = "mistralai/Ministral-3-14B-Reasoning-2512"
adapter_id = "ibitato/c64-ministral-3-14b-thinking-c64-reasoning-lora"
tokenizer = AutoTokenizer.from_pretrained(base_id, trust_remote_code=True)
base_model = Mistral3ForConditionalGeneration.from_pretrained(
base_id,
torch_dtype=torch.bfloat16,
trust_remote_code=True,
)
model = PeftModel.from_pretrained(base_model, adapter_id)
prompt = "Explain the C64 SID chip in one concise paragraph."
inputs = tokenizer(prompt, return_tensors="pt")
with torch.no_grad():
out = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(out[0], skip_special_tokens=True))