| # QModel 6 Configuration Template |
| # ================================== |
| # Copy this to .env and update values for your environment |
| |
| # LLM Backend Selection |
| # Options: "ollama", "hf" (HuggingFace), "gguf" (local GGUF file), or "lmstudio" |
| LLM_BACKEND=ollama |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # OLLAMA BACKEND (if LLM_BACKEND=ollama) |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| OLLAMA_HOST=http://localhost:11434 |
| OLLAMA_MODEL=minimax-m2.7:cloud |
| # Available models: llama3.1, mistral, neural-chat, openhermes |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # HUGGINGFACE BACKEND (if LLM_BACKEND=hf) |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # HF_MODEL_NAME=Qwen/Qwen2-7B-Instruct |
| # HF_DEVICE=auto |
| # HF_MAX_NEW_TOKENS=2048 |
| # Popular models: |
| # - Qwen/Qwen2-7B-Instruct (excellent Arabic) |
| # - mistralai/Mistral-7B-Instruct-v0.2 |
| # - meta-llama/Llama-2-13b-chat-hf |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # GGUF BACKEND (if LLM_BACKEND=gguf) |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # GGUF_MODEL_PATH=./models/Qwen3-32B-Q4_K_M.gguf |
| # GGUF_N_CTX=4096 |
| # GGUF_N_GPU_LAYERS=-1 |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # LM STUDIO BACKEND (if LLM_BACKEND=lmstudio) |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # LMSTUDIO_URL=http://localhost:1234 |
| # LMSTUDIO_MODEL=qwen2.5-7b-instruct |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # EMBEDDING MODEL (shared by all backends) |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| EMBED_MODEL=intfloat/multilingual-e5-large |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # DATA FILES |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| FAISS_INDEX=QModel.index |
| METADATA_FILE=metadata.json |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # RETRIEVAL SETTINGS |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| TOP_K_SEARCH=20 # Candidate pool size |
| TOP_K_RETURN=5 # Final results returned to user |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # GENERATION SETTINGS |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| TEMPERATURE=0.2 # 0.0=deterministic, 1.0=creative |
| MAX_TOKENS=2048 # Max output length |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # SAFETY & QUALITY |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # Confidence threshold: Below this score, skip LLM and return "not found" |
| # Prevents hallucinations but may miss valid results |
| # Range: 0.0-1.0 (default 0.30) |
| # Tune up (0.50+) for stricter, tune down (0.20) for looser |
| CONFIDENCE_THRESHOLD=0.30 |
| |
| # Hadith boost: Score bonus when intent=hadith |
| # Prevents Quran verses from outranking relevant Hadiths |
| HADITH_BOOST=0.08 |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # RANKING |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| RERANK_ALPHA=0.6 # 60% dense (embedding), 40% sparse (BM25) |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # CACHING |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| CACHE_SIZE=512 # Max cache entries |
| CACHE_TTL=3600 # Cache expiry in seconds |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # SECURITY |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| ALLOWED_ORIGINS=* # CORS origins (restrict in production: origin1.com,origin2.com) |
| |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # USAGE EXAMPLES |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ |
| # |
| |
| # LLM_BACKEND=ollama |
| # OLLAMA_HOST=http://localhost:11434 |
| # OLLAMA_MODEL=llama2 |
| # |
| |
| # LLM_BACKEND=hf |
| # HF_MODEL_NAME=Qwen/Qwen2-7B-Instruct |
| # HF_DEVICE=cuda |
| # |
| |
| # LLM_BACKEND=hf |
| # HF_MODEL_NAME=Qwen/Qwen2-7B-Instruct |
| # HF_DEVICE=cpu |
|
|