File size: 1,892 Bytes
301088b 05e66cc 3308b6d 2960ed1 3308b6d b28fb80 e6779cd 3308b6d c85f951 2f86a01 c85f951 3308b6d | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 | ---
title: README
emoji: 📚
colorFrom: pink
colorTo: green
sdk: static
pinned: false
---
**Welcome to Boldt!**
**Boldt** is a family of German language models developed by the **Chair of Machine Learning @ Humboldt-Universität zu Berlin**. This organization hosts our **models, datasets, and research artifacts** related to the Boldt project.
Feel free to explore, download, and experiment with our latest releases! 🚀
## 🌟 The Boldt Model Family
Our models are trained on our German *Dense-Core* subset of FineWeb-2, utilizing a multi-epoch training recipe on high-quality data.
| Model | Parameters | Context Window | Description |
| :--- | :--- | :--- | :--- |
| [**Boldt-DC-350M**](https://huggingface.co/Boldt/Boldt-DC-350M) | 350M | 2048 | Ultra-lightweight base model for constrained environments. |
| [**Boldt-DC-1B**](https://huggingface.co/Boldt/Boldt-DC-1B) | 1B | 2048 | Highly optimized 1B base model with top-tier German performance. |
| [**Boldt-1B**](https://huggingface.co/Boldt/Boldt-1B) | 1B | 4096 | Extended context and augmented with 6B tokens of high-quality German news data. |
| [**Boldt-1B-IT-Preview**](https://huggingface.co/Boldt/Boldt-1B-IT-Preview) | 1B | 4096 | Experimental instruction-tuned model. |
## 📊 Comparison
Boldt-1B compares favorably on German LLM benchmarks against other similarly-sized models:

It is even competitive with many larger (2B parameter) models. See our paper for the full evaluation.
## 📖 Research & Artifacts
* **Paper:** [Repetition over Diversity: High-Signal Data Filtering for Sample-Efficient German Language Modeling (arXiv 2026)](https://arxiv.org/abs/2604.28075)
* **Evaluation Suite:** [Modernized German Benchmarks](https://huggingface.co/collections/Boldt/german-llm-benchmarks)
|