Celeste Imperia | Phi-3.5-mini-instruct (Optimized GGUF)

Highly-optimized GGUF weights for Microsoft's Phi-3.5-mini, specifically forged for high-reasoning tasks on edge hardware.

πŸ› οΈ The Forge

  • Rig: i5-11400 / RTX A4000 / 40GB RAM
  • Context: Validated for 128k long-context execution.

πŸ“‚ Available Precisions

  • FP16 Master: The full-fidelity reference file.
  • Q8_0 Professional: Recommended for complex coding and math.
  • Q4_K_M Mobile: Balanced performance for local agentic workflows.
Downloads last month
16
GGUF
Model size
4B params
Architecture
phi3
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support