--- license: apache-2.0 language: - pt datasets: - AxionLab-Co/ThinkSet-PTBR metrics: - accuracy: 16.9% pipeline_tag: text-generation --- **๐Ÿง  MiniAxion1.5-3M** **Emergent reasoning in a 2.7M parameter model. A tiny Portuguese-first language model that learns how to think before it learns how to be correct.** **๐Ÿš€ Overview** MiniAxion1.5-3M is an ultra-compact (~2.7M parameters) GPT-style language model designed to investigate reasoning emergence at extreme small scale. Unlike typical small models optimized for fluency, MiniAxion is explicitly trained to produce: Structured reasoning traces Step-by-step thinking () Deterministic answer formatting It operates primarily in Portuguese, making it a rare example of a non-English reasoning-first nano model. **โšก Why This Model Is Interesting** Most models follow this trajectory: Language โ†’ Knowledge โ†’ Reasoning MiniAxion flips part of that: Structure โ†’ Reasoning format โ†’ (still learning correctness) **๐Ÿ’ก Key insight:** The model demonstrates that reasoning structure can emerge independently of reasoning accuracy. **๐Ÿงช Evaluation** Task Performance Task Accuracy Addition 10% Subtraction 10% Multiplication 0% Even/Odd 100% Comparison 5% Sequence Completion 0% Word Problems (Addition) 10% Word Problems (Subtraction) 0% Word Problems (Multiplication) 10% True/False 100% Chat/Greetings 100% **๐Ÿง  Reasoning Behavior Metrics** Metric Score Thinking Rate 100% Step Format 100% Answer Completion 100% โœ” The model always thinks โœ” The model always structures reasoning โœ” The model always produces an answer **๐Ÿ“Š Interpretation** MiniAxion exhibits a clear dissociation: โœ… What it learned Reasoning format Step-by-step decomposition Logical task patterns (parity, boolean) โŒ What it did NOT learn Arithmetic correctness Numerical reasoning Multi-step computation **๐Ÿ”ฌ Core Finding** Reasoning โ‰  Correctness MiniAxion shows that: Models can internalize thinking patterns Without actually learning how to solve problems This makes it a strong candidate for studying: Emergent reasoning Tiny Recursive Models (TRMs) Reasoning distillation **๐Ÿ—๏ธ Architecture** Type: GPT-style Transformer Parameters: ~2.7M Objective: Next-token prediction Language: Portuguese (primary) Specialization: Structured reasoning traces **๐Ÿง  Training Strategy** The model was trained with a reasoning-first approach: Portuguese language grounding Structured reasoning data () Emphasis on: Deterministic formats Multi-step thinking Explicit reasoning tokens ๐Ÿšซ No RLHF ๐Ÿšซ No instruction tuning at scale ๐Ÿšซ No large model distillation (yet) โš ๏ธ Limitations 1. Arithmetic Collapse Near-random performance in: Addition Subtraction Multiplication โ†’ Indicates lack of numerical representation learning Strong dependence on: Prompt format Token patterns Seen reasoning templates **๐Ÿ”ฎ Future Work** This model is just the beginning. ๐Ÿ“ˆ Scaling 5M / 10M / 20M versions Track emergence of correctness ๐Ÿงช Distillation Inject reasoning from larger models Improve accuracy without scaling params ๐Ÿ” Self-Play / Synthetic Data Generate reasoning loops Reinforce correct chains ๐Ÿงฉ Hybrid Reasoning Combine symbolic + neural learning Fix arithmetic weakness ๐Ÿงพ Example Output Identifico os nรบmeros Tento somar os valores Ajusto o resultado 74 โœ” Perfect reasoning structure โŒ Incorrect answer **๐Ÿ’ก Takeaway** MiniAxion1.5-3M proves something important: Even a 2.7M model can learn to simulate thinking before it learns to actually think correctly. **๐Ÿค Use Cases** Research on emergent reasoning Tiny model experimentation (CPU-friendly) Educational demos of: Chain-of-Thought Reasoning failure modes Base model for: Distillation NRM experiments