avara-x1-mini
Collection
All avara-x1-mini models • 2 items • Updated
Avara X1 Mini is a lightweight AI model developed by Omnionix. This version is provided in GGUF format, optimized for local inference on CPUs, GPUs, and mobile devices. Based on the Qwen2.5 architecture, it balances technical reasoning with a grounded and supportive personality.
Join the Community: Omnionix Discord
| Feature | Details |
|---|---|
| Developer | Omnionix |
| Architecture | Qwen2.5-1.5B |
| Format | GGUF (Static Quantization) |
| Prompt Template | ChatML |
| Recommended for | Mobile, Low-RAM PC, Edge devices |
These GGUF files were created using llama.cpp. We recommend Q4_K_M for the best balance of speed and intelligence, or Q8_0 for near-original performance.
To run Avara X1 Mini locally, you can use LM Studio, Ollama, or llama-cpp-python.
from llama_cpp import Llama
llm = Llama(
model_path="./avara-x1-mini-q4_k_m.gguf",
n_ctx=2048,
n_threads=4,
chat_format="chatml"
)
print("\n--- Avara X1 Mini (GGUF) is Online ---")
response = llm.create_chat_completion(
messages = [
{"role": "system", "content": "You are Avara, an AI assistant created by Omnionix."},
{"role": "user", "content": "Hello Avara, what can you do?"}
]
)
print(response["choices"][0]["message"]["content"])
4-bit