Llama-3.2-3B-Instruct GGUF (Q4_K_M)

Personally quantised GGUF version of meta-llama/Llama-3.2-3B-Instruct, quantised using llama.cpp.


πŸŽ“ About OpenMarker

This model was quantised primarily for use with OpenMarker β€” an open source AI assistant designed to help professors, lecturers, and educators automatically grade and review student work.

OpenMarker uses local quantised models so that:

  • No student data is sent to external APIs
  • It runs fully offline and privately
  • Anyone can self-host it for free

πŸ‘‰ https://github.com/theDALEX/openmarker


πŸ“¦ Quant Details

Quant Size Description
Q4_K_M ~2GB Good quality, recommended for most use cases

Quantised using llama.cpp on Windows.


πŸ”— Original Model

  • Original weights: meta-llama/Llama-3.2-3B-Instruct
  • Original license: Llama 3.2 Community License
  • Changes made: None β€” weights are unchanged beyond quantisation
Downloads last month
286
GGUF
Model size
3B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for dalexdavis/Llama-3.2-3B-Instruct-Q4_K_M-GGUF

Quantized
(439)
this model