Avara X1 Mini Logo


Avara X1 Mini (GGUF)

Avara X1 Mini is a lightweight AI model developed by Omnionix. This version is provided in GGUF format, optimized for local inference on CPUs, GPUs, and mobile devices. Based on the Qwen2.5 architecture, it balances technical reasoning with a grounded and supportive personality.

Join the Community: Omnionix Discord


Technical Specifications

Feature Details
Developer Omnionix
Architecture Qwen2.5-1.5B
Format GGUF (Static Quantization)
Prompt Template ChatML
Recommended for Mobile, Low-RAM PC, Edge devices

Quantization Info

These GGUF files were created using llama.cpp. We recommend Q4_K_M for the best balance of speed and intelligence, or Q8_0 for near-original performance.


Implementation

To run Avara X1 Mini locally, you can use LM Studio, Ollama, or llama-cpp-python.

Using llama-cpp-python

from llama_cpp import Llama

llm = Llama(
    model_path="./avara-x1-mini-q4_k_m.gguf",
    n_ctx=2048,
    n_threads=4,
    chat_format="chatml"
)

print("\n--- Avara X1 Mini (GGUF) is Online ---")

response = llm.create_chat_completion(
    messages = [
        {"role": "system", "content": "You are Avara, an AI assistant created by Omnionix."},
        {"role": "user", "content": "Hello Avara, what can you do?"}
    ]
)

print(response["choices"][0]["message"]["content"])
Downloads last month
35
GGUF
Model size
2B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including OmnionixAI/avara-x1-mini-Q4_K_M-GGUF