LetheanNetwork/lemer-mlx

Gemma 4 E2B in MLX format, 4-bit quantized, converted from LetheanNetwork/lemer's bf16 safetensors via mlx_lm.convert. This is the unmodified Google Gemma 4 E2B-IT weights — no LEK shift, no fine-tuning — hosted in our namespace so downstream tools (benchmarks, apps) don't have to depend on external mlx-community mirrors.

For the LEK-merged (consent-based ethical kernel) variant of the same model, see lthn/lemer.

Variants in this family

Repo Format Bits Use case
LetheanNetwork/lemer safetensors + gguf Q4_K_M bf16 / 4 Source weights + llama.cpp/Ollama
LetheanNetwork/lemer-mlx mlx 4 This repo — Apple Silicon default
LetheanNetwork/lemer-mlx-8bit mlx 8 Apple Silicon higher-precision
LetheanNetwork/lemer-mlx-bf16 mlx bf16 Apple Silicon full-precision reference

Usage

from mlx_lm import load, generate

model, tokenizer = load("LetheanNetwork/lemer-mlx")
response = generate(
    model, tokenizer,
    prompt=tokenizer.apply_chat_template(
        [{"role": "user", "content": "Hello"}],
        add_generation_prompt=True,
        enable_thinking=True,
    ),
    max_tokens=512,
)

Provenance

  • Source: LetheanNetwork/lemer bf16 safetensors (= google/gemma-4-E2B-it)
  • Converter: mlx_lm.convert (mlx-lm — LM Studio / Apple ML Research)
  • Quant: 4-bit group quantization, ~4.5 bits/weight effective
  • License: Apache 2.0 (Gemma Terms of Use)

License

Apache 2.0, subject to the Gemma Terms of Use.

Downloads last month
52
Safetensors
Model size
0.7B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LetheanNetwork/lemer-mlx

Quantized
(108)
this model