QVikhr-3-8B-Instruction-f32-GGUF

QVikhr-3-8B-Instruction is a powerful bilingual instruction-tuned language model based on the Qwen3-8B architecture, further refined with supervised fine-tuning (SFT) on the GrandMaster2 dataset, which is notable for its focus on Russian and English language data. This targeted training enables the model to generate precise, context-sensitive responses for a wide array of instruction-based tasks, such as answering questions, following detailed directions, and analyzing texts—particularly excelling at complex topics like mathematics and physics in Russian. By leveraging 8 billion parameters and advanced instructional alignment, QVikhr-3-8B-Instruction delivers near state-of-the-art results, matching or approaching the performance of larger models like GPT-4.1 on competitive benchmarks, while maintaining efficiency and versatility suitable for integration in both professional and user-facing applications.

Execute using Ollama

run ->

ollama run hf.co/prithivMLmods/QVikhr-3-8B-Instruction-f32-GGUF:BF16

Model Files

File Name Quant Type File Size
QVikhr-3-8B-Instruction.BF16.gguf BF16 16.4 GB
QVikhr-3-8B-Instruction.F16.gguf F16 16.4 GB
QVikhr-3-8B-Instruction.Q8_0.gguf Q8_0 8.71 GB

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
36
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/QVikhr-3-8B-Instruction-f32-GGUF

Finetuned
Qwen/Qwen3-8B
Quantized
(6)
this model