AMD-HybridLM-Models ✨
Collection
AMD-HybridLM is a family of post-trained, highly efficient hybrid models, designed to combine performance with speed and memory efficiency. • 21 items • Updated • 5
This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct on the JunxiongWang/sftdatasetv3, the amd/OpenMathInstruct-2_ZebraLlama_2M, the amd/Zebra_Llama_OpenThoughts-114k-math, the amd/OpenR1-Math-220K and the amd/Zebra_Llama_ChatQA2-Long-SFT_long_sft_QA datasets. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 1522.4293 | 1.0 | 38873 | 1436.5125 |
Base model
meta-llama/Llama-3.2-1B-Instruct