This is just my IQ4_XS quant of https://huggingface.co/failspy/Meta-Llama-3-70B-Instruct-abliterated-v3.5
Also included is my generated imatrix data file, generated on 200 x 2048 tokens of Wiki.Train.Raw against Q8_0 (i dont have enough RAM + VRAM to do the full fp16).
- Downloads last month
- 19
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support