Safetensors
English
qwen3_5

Aureth Qwen3.5‑0.8B

OusiaResearch / Aureth_Qwen3.5‑0.8B — a lean, efficient fine-tune of Qwen3.5‑0.8B, purpose-built for light inference workloads and resource‑constrained deployments.


∴ Model Summary

Field Detail
Base model Qwen/Qwen3.5‑0.8B
Parameters 0.9 B (full precision)
Context length 4 096 tokens
Training volume 65 M tokens
Precision BF16 safetensors + FP32
License Apache 2.0
Framework Unsloth + HuggingFace TRL

Aureth — from the Ouroboros axiom ουσία (ousia, "being" or "essence") — denotes a model stripped to its functional core: efficient, present, and ready to run where larger models cannot.


◦ Training

Corpus

Single‑phase full fine‑tuning on a curated mix of:

  • Synthetic cognitive traces — Hermes‑4.3‑36B‑generated reasoning chains
  • Instructional dialogue — structured for clarity and concision
  • Domain‑neutral conversational data — broad coverage over specialist depth

The 65 M token volume is intentionally bounded — calibrated for lightweight adaptation without memorisation or alignment degradation.

Objective

Standard next‑token prediction (cross‑entropy) at full precision (bf16 → fp32) — no adapter layers, no quantisation. The full parameter space is updated at modest learning rate to preserve the base model's innate capabilities while gently steering tone and instruction‑following.

Hardware

Trained via Unsloth Studio — 2× faster than naive bf16 fine‑tuning with equivalent or better loss convergence.


○ Intended Use

 lightest workloads · embedded hosts · edge devices
 rapid prototyping · 24‑7 low‑latency inference

Aureth‑0.8B is not a reasoning model. It excels at:

  • Simple Q&A and information retrieval
  • Short‑context summarisation
  • Light conversational agents
  • Scripted automation / tool‑calling at low branching depth
  • Running as a always‑on personal assistant on modest hardware (MacBook, single‑GPU workstation, CPU inference via llama.cpp/MLX)

◦ Ousia Research

Aureth is the production fine‑tune lineage of the Ousia Project — a family of models trained on synthetically generated cognitive traces with the goal of developing coherent, consistent, and ethically grounded language models.

Variant Context Corpus Status
Aureth‑Qwen3.5‑0.8B 4 K 65 M tokens ◉ Public
Aureth‑Qwen3.5‑4B 4 K ~262 M tokens ∙ Training
Aureth‑Qwen3.5‑9B 4 K TBD ∙ Planned

"The smallest Ouroboros still eats its own tail."


Model card v1.0 — Ousia Research — April 2026 Licensed under Apache 2.0 — OusiaResearch / Aureth_Qwen3.5‑0.8B

Downloads last month
50
Safetensors
Model size
0.9B params
Tensor type
F32
·
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for OusiaResearch/Aureth_Qwen3.5-0.8B

Finetuned
(177)
this model
Quantizations
2 models

Datasets used to train OusiaResearch/Aureth_Qwen3.5-0.8B