Fix AI Analyst: use Qwen2.5-7B as default HF model
Browse filesRayMelius/stockex-analyst is not hosted on any HF inference provider,
causing HTTP 400 errors. Fall back to Qwen/Qwen2.5-7B-Instruct.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- ai_analyst/ai_analyst.py +1 -1
ai_analyst/ai_analyst.py
CHANGED
|
@@ -11,7 +11,7 @@ from shared.kafka_utils import create_producer, create_consumer
|
|
| 11 |
OLLAMA_HOST = os.getenv("OLLAMA_HOST", "") # e.g. http://host.docker.internal:11434
|
| 12 |
OLLAMA_MODEL = os.getenv("OLLAMA_MODEL", "llama3.1:8b")
|
| 13 |
HF_TOKEN = os.getenv("HF_TOKEN", "")
|
| 14 |
-
HF_MODEL = os.getenv("HF_MODEL", "
|
| 15 |
GROQ_API_KEY = os.getenv("GROQ_API_KEY", "")
|
| 16 |
GROQ_MODEL = os.getenv("GROQ_MODEL", "llama-3.1-8b-instant")
|
| 17 |
GROQ_URL = "https://api.groq.com/openai/v1/chat/completions"
|
|
|
|
| 11 |
OLLAMA_HOST = os.getenv("OLLAMA_HOST", "") # e.g. http://host.docker.internal:11434
|
| 12 |
OLLAMA_MODEL = os.getenv("OLLAMA_MODEL", "llama3.1:8b")
|
| 13 |
HF_TOKEN = os.getenv("HF_TOKEN", "")
|
| 14 |
+
HF_MODEL = os.getenv("HF_MODEL", "Qwen/Qwen2.5-7B-Instruct")
|
| 15 |
GROQ_API_KEY = os.getenv("GROQ_API_KEY", "")
|
| 16 |
GROQ_MODEL = os.getenv("GROQ_MODEL", "llama-3.1-8b-instant")
|
| 17 |
GROQ_URL = "https://api.groq.com/openai/v1/chat/completions"
|