Post
48
I taught an old dog, or in this case model, new tricks. Meet Galactic Reasoning 1.3B: https://huggingface.co/collections/TitleOS/galactic-reasoning-galactica-with-chain-of-thought. By finetuning Meta's (at the time Facebook) Galactica model against
glaiveai glaiveai/reasoning-v1-20m. After training for 1000 steps on my poor overworked Tesla P40 for 48 hours, I was able to produce a merged FP16, LoRA and quantization Q8 weights. Check out the readme.md for an example CoT.