Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
+
tags:
|
| 5 |
+
- bar
|
| 6 |
+
- mixture-of-experts
|
| 7 |
+
- olmo
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
# BAR
|
| 11 |
+
|
| 12 |
+
BAR (Branch-Adapt-Route) is a modular post-training approach that extends a fully post-trained language model with new domain capabilities via independently trained Mixture-of-Experts. Rather than retraining a single model across all domains, BAR trains independent domain experts β each through its own mid-training, supervised finetuning (SFT), and reinforcement learning pipeline β and composes them into a unified model via an MoE architecture with lightweight router training.
|
| 13 |
+
|
| 14 |
+
All BAR models are built on top of Olmo 2 7B.
|
| 15 |
+
|
| 16 |
+
## Models in the BAR suite
|
| 17 |
+
|
| 18 |
+
- **BAR-7B** β initial fully post-trained 7B dense model (the starting point)
|
| 19 |
+
- **BAR-2x7B-Base** β 2-expert MoE (anchor + base pre-trained model)
|
| 20 |
+
- **BAR-2x7B-Math-SFT** β math expert after mid-training and SFT
|
| 21 |
+
- **BAR-2x7B-Math** β math expert after mid-training + SFT + RLVR
|
| 22 |
+
- **BAR-2x7B-Code-SFT** β code expert after mid-training and SFT
|
| 23 |
+
- **BAR-2x7B-Code** β code expert after mid-training + SFT + RLVR
|
| 24 |
+
- **BAR-2x7B-Tool-Use** β tool use expert (SFT only)
|
| 25 |
+
- **BAR-2x7B-Safety** β safety expert (SFT only)
|
| 26 |
+
- **BAR-5x7B** β final 5-expert MoE combining all experts with a trained router
|
| 27 |
+
|
| 28 |
+
## Results
|
| 29 |
+
|
| 30 |
+
| Model | Overall | Knowledge | Reasoning | Chat | Math | Code | Tool Use | Safety |
|
| 31 |
+
|---|---|---|---|---|---|---|---|---|
|
| 32 |
+
| BAR-7B | 31.3 | 28.5 | 29.8 | 48.9 | 23.6 | 11.8 | 25.3 | 51.3 |
|
| 33 |
+
| BAR-2x7B-Math-SFT | 36.8 | 28.8 | 31.2 | 40.9 | 41.9 | 20.5 | 21.6 | 72.7 |
|
| 34 |
+
| BAR-2x7B-Math | 39.3 | 29.0 | 30.8 | 42.5 | 55.8 | 22.1 | 19.8 | 75.4 |
|
| 35 |
+
| BAR-2x7B-Code-SFT | 38.5 | 28.8 | 29.1 | 40.1 | 25.5 | 49.3 | 19.7 | 77.3 |
|
| 36 |
+
| BAR-2x7B-Code | 38.8 | 28.5 | 29.2 | 41.0 | 26.9 | 50.4 | 19.8 | 75.3 |
|
| 37 |
+
| BAR-2x7B-Tool-Use | 37.2 | 28.5 | 28.7 | 39.3 | 21.8 | 16.9 | 46.4 | 79.1 |
|
| 38 |
+
| BAR-2x7B-Safety | 35.6 | 28.7 | 28.8 | 38.1 | 22.4 | 15.7 | 21.1 | 94.6 |
|
| 39 |
+
| BAR-5x7B | 49.1 | 28.4 | 30.8 | 38.7 | 56.2 | 49.9 | 45.6 | 94.0 |
|
| 40 |
+
|
| 41 |
+
Scores are unweighted averages across benchmarks within each category. See the paper for per-benchmark results and full evaluation details.
|
| 42 |
+
|
| 43 |
+
## License
|
| 44 |
+
|
| 45 |
+
This model is licensed under Apache 2.0. It is intended for research and educational use in accordance with Ai2's [Responsible Use Guidelines](https://allenai.org/responsible-use).
|