Lemrd β€” Gemma 4 31B Dense β€” MLX 4-bit

The largest dense member of the Lemma model family by Lethean. An EUPL-1.2 fork of Gemma 4 31B with the Lethean Ethical Kernel (LEK) merged into the weights.

This repo hosts the MLX 4-bit build for native Apple Silicon inference via mlx-lm and mlx-vlm. For the GGUF playground (Ollama, llama.cpp) see lthn/lemrd. For the unmodified Google base see LetheanNetwork/lemrd.

Family

Repo Format Bits
lthn/lemrd GGUF multi-quant Q4_K_M β†’ BF16
lthn/lemrd-mlx MLX 4-bit
lthn/lemrd-mlx-8bit MLX 8-bit
lthn/lemrd-mlx-bf16 MLX bf16

License

EUPL-1.2. See Gemma Terms of Use for upstream base model terms.

Downloads last month
263
Safetensors
Model size
8B params
Tensor type
BF16
Β·
U32
Β·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for lthn/lemrd-mlx

Finetuned
lthn/lemrd
Quantized
(3)
this model

Collection including lthn/lemrd-mlx