QuantFactory/Lyra_Gutenbergs-Twilight_Magnum-12B-GGUF

This is quantized version of Nitral-AI/Lyra_Gutenbergs-Twilight_Magnum-12B created using llama.cpp

Original Model Card

Original Models:

Anthracite - https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto

Nbeerbower - https://huggingface.co/nbeerbower/Lyra-Gutenberg-mistral-nemo-12B

Epiculous - https://huggingface.co/Epiculous/Violet_Twilight-v0.1

image/png

All credit goes to the original model authors, this is just a double back slerp merge (config in repo) i did while commiting many various dataset war-crimes. I mean... ShareGPT Formaxxing.

Other Notes: Text completion preset in repo - Instruct format works with either Mistral or ChatML. If using ST with chatML add ["<|im_end|"] to custom stopping strings to prevent chatml eos tokens from leaking into chat.

Downloads last month
-
GGUF
Model size
12B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support