Leonidas-4B

"Come and take them." โ€” Leonidas I, 480 BC

Few parameters. No retreat.


Model Description

Leonidas-4B is a fine-tuned Polish reasoning model built on the Qwen3.5-4B hybrid architecture (Mamba + Attention). It was trained using LoRA fp16 on a curated 48k Polish Chain-of-Thought dataset, with native <think> reasoning blocks.

Forged by Hattori Hanzo โ€” because an idiot admires complexity, a genius admires simplicity.

"An idiot admires complexity, a genius admires simplicity." โ€” Terry A. Davis, TempleOS


Architecture

Property Value
Base Model Qwen3.5-4B (hybrid Mamba+Attention)
Parameters ~4B
Training Method LoRA fp16 (r=16, alpha=16)
Trainable params 0.47%
Training Steps 1500
Final Loss 0.4389
Context Length 32768
Language Polish ๐Ÿ‡ต๐Ÿ‡ฑ + English

Training

  • Dataset: 48k Polish CoT (Chain-of-Thought) โ€” mixed reasoning, math, logic, science
  • Format: ChatML with native <think> blocks
  • Platform: Kaggle T4 16GB
  • Phases:
    Phase Steps LR Scheduler
    1 2000 2e-4 linear
    2 3000 5e-5 cosine
    3 5000 3e-5 cosine
    4 5000 1e-5 cosine

Available Files

File Size Description
leonidas_f16.gguf 8.41 GB Full precision fp16
leonidas_Q8_0.gguf 4.47 GB Q8_0 โ€” best quality
leonidas_Q6_K.gguf 3.45 GB Q6_K โ€” great quality
leonidas_Q4_K_M.gguf 2.70 GB Q4_K_M โ€” recommended for most setups

Usage

llama.cpp

./llama-cli \
  -m leonidas_Q4_K_M.gguf \
  -p "Jesteล› asystentem AI. Myล›l krok po kroku." \
  --chat-template chatml \
  -n 1024

Ollama / OpenWebUI

Compatible with any OpenAI-compatible frontend supporting GGUF + ChatML template. ### Recommended Parameters temperature: 0.6 top_p: 0.95 top_k: 40 repetition_penalty: 1.05

Capabilities

  • โœ… Native Polish reasoning with <think> CoT blocks
  • โœ… Mathematics and logic
  • โœ… Scientific explanations
  • โœ… Sentiment analysis
  • โœ… Code generation (Python, basics)
  • โœ… Quantum physics concepts

Forge Stamp

general.author:   HattoriHanzo1
hanzo.base_model: Qwen3.5-4B
hanzo.leonidas:   ฮคฯŒฯ„ฮต แผฮฝ ฯ„แฟ‡ ฯƒฮบฮนแพท ฮผฮฑฯ‡ฮฟฯฮผฮตฮธฮฑ
                  ("Then we shall fight in the shade.")

License

Apache 2.0 โ€” base model license from Qwen3.5-4B (Alibaba Cloud).

Romani ite domum ๐Ÿ˜„ โ€” Qapla'! โš”๏ธ๐Ÿ––

Downloads last month
1,493
GGUF
Model size
4B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

4-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for HattoriHanzo1/Leonidas-4B-GGUF

Finetuned
Qwen/Qwen3.5-4B
Adapter
(100)
this model