Gemma 4 E4B — AI Job Searcher (GGUF Q4_K_M)

Fine-tuned google/gemma-4-E4B-it for multilingual job search assistance. Quantized to Q4_K_M for efficient local inference via llama.cpp.

Model Details

Property Value
Base model google/gemma-4-E4B-it
Fine-tune method LoRA (r=16, α=16)
Training hardware NVIDIA RTX 5080 (16GB VRAM)
Quantization Q4_K_M (~5 GB)
Format GGUF (llama.cpp compatible)
Languages EN, ES, FR, DE, PT, NO, DA, FI, SV
Task Job search assistance, CV help, interview prep

Training

LoRA Adapter

The unmerged LoRA adapter is available at: ai-colombia/gemma4-e4b-job-searcher-lora

Usage

llama.cpp / LM Studio / Ollama

./llama-cli -m gemma4-e4b-job-searcher-q4_k_m.gguf --chat-template gemma -p "You are a helpful job search assistant." -i

Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(
    model_path="gemma4-e4b-job-searcher-q4_k_m.gguf",
    n_ctx=2048,
    n_gpu_layers=-1,  # use GPU if available
)

response = llm.create_chat_completion(
    messages=[
        {"role": "system", "content": "You are a helpful job search assistant."},
        {"role": "user", "content": "Help me write a cover letter for a software engineer position."}
    ]
)
print(response["choices"][0]["message"]["content"])

Capabilities

  • Job search guidance — advice on finding jobs, job boards, networking
  • CV / Resume writing — structure, content, ATS optimization tips
  • Cover letter writing — tailored letters for specific roles
  • Interview preparation — common questions, STAR method, salary negotiation
  • Career advice — career transitions, skill gaps, industry insights
  • Multilingual — responds in EN, ES, FR, DE, PT, NO, DA, FI, SV

Limitations

  • Based on google/gemma-4-E4B-it — subject to Gemma's usage policy
  • Knowledge cutoff from base model training data
  • Q4_K_M quantization may reduce quality on complex reasoning vs the full fp16 model
  • Not suitable for real-time job listings (no web access)

License

This model is subject to the Gemma Terms of Use. The fine-tuning data and LoRA weights are released under Apache 2.0.

Downloads last month
292
GGUF
Model size
8B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ai-colombia/gemma4-e4b-job-searcher-gguf

Adapter
(54)
this model