gemma-2-9B-it-iq1_m
This is a quantized version of the Gemma2 9B instruct model using the IQ1_M quantization method.
Model Details
- Original Model: Gemma2-9B-it
- Quantization Method: IQ1_M
- Precision: 1-bit
- iMatrix: From bartowski. You can find the file in the following repo; gemma-2-9b-it-gguf repo
Usage
You can use it directly with llama.cpp
- Downloads last month
- 7
Hardware compatibility
Log In to add your hardware
1-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support