Hours-1.0-4B-GGUF

Horus Model

GGUF quantized versions of Horus-1.0-4B by TokenAI.

Base Model

  • Source: tokenaii/horus
  • Original Model: Horus-1.0-4B (4B parameters)
  • Developer: Assem Sabry & TokenAI
  • Organization: TokenAI
  • Release Date: April 2026
  • License: MIT

About TokenAI

TokenAI is an AI startup founded by Assem Sabry with headquarters in Egypt.

Mission

TokenAI aims to deliver the strongest language models in the world and in the Arab world through the Horus family of models. The startup bridges the gap between cutting-edge AI capabilities and regional cultural contexts, starting with the Arab world.

The Horus Family

Horus-1.0-4B marks the first model in the Horus family line. This is just the beginning of TokenAI's journey to create a comprehensive suite of AI models serving the Arab region.

Horus-1.0-4B-GGUF

GGUF quantized versions of Horus-1.0-4B - A 4B parameter multilingual language model optimized for Arabic and English.

Model Variants & Hardware Requirements

Format File Size Min RAM (CPU) Min VRAM (GPU) Quality Best For
F16 9.03 GB 12 GB 10 GB Maximum quality High-end GPUs (RTX 3090, A100)
Q8_0 4.8 GB 6 GB 5 GB Near-lossless RTX 3060 12GB, RTX 4060
Q6_K 3.71 GB 5 GB 4 GB Excellent RTX 3060, RTX 4060 Laptop
Q5_K_M 3.23 GB 4 GB 3.5 GB Very Good GTX 1650, RTX 3050
Q4_K_M 2.78 GB 3.5 GB 3 GB Good Entry-level GPUs, CPU-only

Detailed Hardware Requirements

F16 (FP16 - Full Precision)

  • File: Horus-1.0-4B-F16.gguf (9.03 GB)
  • Min System RAM: 12 GB
  • Min VRAM: 10 GB
  • Recommended: RTX 3090, RTX 4090, A100, A6000
  • Use Case: Maximum quality, research, fine-tuning reference

Q8_0 (8-bit Quantization)

  • File: Horus-1.0-4B-Q8_0.gguf (4.8 GB)
  • Min System RAM: 6 GB
  • Min VRAM: 5 GB
  • Recommended: RTX 3060 12GB, RTX 4060, RTX 4070
  • Use Case: Near-lossless quality with half the memory

Q6_K (6-bit K-Quant)

  • File: Horus-1.0-4B-Q6_K.gguf (3.71 GB)
  • Min System RAM: 5 GB
  • Min VRAM: 4 GB
  • Recommended: RTX 3060, RTX 4060 Laptop, GTX 1080 Ti
  • Use Case: Excellent quality for most applications

Q5_K_M (5-bit K-Quant Medium)

  • File: Horus-1.0-4B-Q5_K_M.gguf (3.23 GB)
  • Min System RAM: 4 GB
  • Min VRAM: 3.5 GB
  • Recommended: GTX 1650 Super, RTX 3050, RTX 3050 Ti
  • Use Case: Balanced quality and performance

Q4_K_M (4-bit K-Quant Medium)

  • File: Horus-1.0-4B-Q4_K_M.gguf (2.78 GB)
  • Min System RAM: 3.5 GB
  • Min VRAM: 3 GB
  • Recommended: GTX 1060 6GB, GTX 1650, Intel Arc A380
  • Use Case: Maximum compression, edge devices, CPU inference

Quick Start

Using NeuralNode (Recommended)

The easiest way to use Horus GGUF models is with the NeuralNode framework:

import neuralnode as nn

MODEL_ID = "tokenaii/Hours-1.0-4B-GGUF/Horus-1.0-4B-Q6_K.gguf"
DEVICE = "cpu"  # Change to "cuda" for GPU acceleration

# Download and load
model = nn.HorusModel(MODEL_ID, device=DEVICE).load()

# Use immediately
response = model.chat([{"role": "user", "content": "hi horus im emy"}])
print(response.content)

Using llama-cpp-python

For direct llama.cpp integration:

from llama_cpp import Llama

llm = Llama(
    model_path="Horus-1.0-4B-Q4_K_M.gguf",
    n_ctx=4096
)

output = llm("Hello, how are you?", max_tokens=256)
print(output['choices'][0]['text'])

Voice Interface with Replica TTS

Add natural voice output to your Horus GGUF model with Replica TTS:

import neuralnode as nn

voice_id = "replica-aria-language{en-us}"

MODEL_ID = "tokenaii/Hours-1.0-4B-GGUF/Horus-1.0-4B-F16.gguf"
DEVICE = "cuda"

# Load model with Replica TTS
model = nn.HorusModel(
    MODEL_ID,
    tts_engine="replica_tts",
    voice=voice_id,
    device=DEVICE
).load()

# Chat and get spoken response
response = model.chat([{"role": "user", "content": "Hello!"}])
print(response.content)
response.play_audio()  # Plays the TTS audio

Browse All Voices

import neuralnode as nn

voices = nn.replica_voice_list()
for voice in voices:
    print(voice)

Benchmark Results

Below are visual comparisons of Horus-1.0-4B against leading models.

General Knowledge & Reasoning

General Benchmarks

Arabic Language & Cultural Benchmarks

Arabic Benchmarks

Coding & Tool Use Benchmarks

Coding Benchmarks


Model Capabilities

  • Multilingual: Supports 10+ languages including Arabic, English, French, Spanish, German, Italian, Portuguese, Turkish, Urdu, Hindi
  • Identity Recognition: Knows itself as Horus from TokenAI
  • Reasoning: Chain-of-thought capabilities
  • Context Length: Up to 4096 tokens
  • Voice Output: Replica TTS integration for natural speech

Links


Note: Quantized using llama.cpp for efficient inference. GGUF versions are optimized for local deployment with minimal resource requirements.

Downloads last month
1,006
GGUF
Model size
5B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tokenaii/Horus-1.0-4B-GGUF

Base model

tokenaii/horus
Quantized
(1)
this model