tf3-26m-student / README.md
mihainadas's picture
Upload README.md with huggingface_hub
cf1c14c verified
---
license: apache-2.0
language:
- ro
library_name: transformers
pipeline_tag: text-generation
tags:
- llama
- romanian
- synthetic-data
- distillation
- tinyfabulist
- fables
base_model: klusai/tf3-50m-base
datasets:
- klusai/ds-tf2-en-ro-15k
---
# TF3 Student: Distilled Romanian Language Model
A compact **22.9M-parameter** Romanian language model distilled from the [TF3-50M teacher](https://huggingface.co/klusai/tf3-50m-base) using logit-based knowledge distillation. Part of the [TinyFabulist](https://arxiv.org/abs/2601.10410) research project.
## Model Details
| Property | Value |
|----------|-------|
| Parameters | 22.9M (26.45M with untied embeddings) |
| Architecture | LLaMA-style decoder-only Transformer |
| Hidden size | 384 |
| Attention heads | 6 (head dim 64) |
| Layers | 6 |
| MLP intermediate | 1,024 |
| Vocab size | 32,000 (Unigram, Romanian-specific) |
| Context length | 2,048 tokens |
| Tied embeddings | Yes |
| Training | Knowledge distillation from klusai/tf3-50m-base |
## Training
- **Method**: Logit-based knowledge distillation (KL + CE loss, alpha=0.009)
- **Teacher**: [klusai/tf3-50m-base](https://huggingface.co/klusai/tf3-50m-base) (51.65M params, frozen)
- **Data**: [klusai/ds-tf2-en-ro-15k](https://huggingface.co/datasets/klusai/ds-tf2-en-ro-15k) (15k Romanian fables)
- **Temperature**: T=1.0
- **Epochs**: 3
- **Learning rate**: 3e-4 (cosine schedule, 50-step warmup)
- **Hardware**: Apple M3 Ultra (96GB unified memory)
## Intended Use
This model is a research artifact demonstrating knowledge distillation for compact Romanian language models trained on synthetic moral microfiction. It is designed for:
- Research on compact language model compression
- Romanian text generation in the fable/moral story domain
- Downstream fine-tuning for Romanian NLP tasks
**Not intended for**: Production text generation, factual question answering, or safety-critical applications.
## Limitations
- Domain-restricted to moral microfiction (fables)
- Trained exclusively on synthetic data
- May exhibit repetitive patterns and simplified phrasing compared to the teacher
- Gender agreement errors may occur in generated text
## Citation
```bibtex
@article{nadas2026tf3,
title={TF3-RO-50M: Training Compact Romanian Language Models from Scratch on Synthetic Moral Microfiction},
author={Nada\c{s}, Mihai Dan and Dio\c{s}an, Laura and Tomescu, Andreea and Pi\c{s}coran, Andrei},
journal={arXiv preprint arXiv:2601.10410},
year={2026}
}
```
## Related Models and Datasets
| Artifact | Description |
|----------|-------------|
| [klusai/tf3-50m-base](https://huggingface.co/klusai/tf3-50m-base) | Teacher model (51.65M) |
| [klusai/tf3-50m-sft](https://huggingface.co/klusai/tf3-50m-sft) | SFT-tuned teacher |
| [klusai/tf3-bert](https://huggingface.co/klusai/tf3-bert) | NER model for entity coherence evaluation |
| [klusai/ds-tf2-en-ro-3m](https://huggingface.co/datasets/klusai/ds-tf2-en-ro-3m) | 3M bilingual fable corpus |
| [klusai/ds-tf2-en-ro-15k](https://huggingface.co/datasets/klusai/ds-tf2-en-ro-15k) | 15k curated subset for distillation/SFT |