RayMelius Claude Opus 4.6 commited on
Commit
7559e0e
·
1 Parent(s): 3326378

Fix AI Analyst: use Qwen2.5-7B as default HF model

Browse files

RayMelius/stockex-analyst is not hosted on any HF inference provider,
causing HTTP 400 errors. Fall back to Qwen/Qwen2.5-7B-Instruct.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>

Files changed (1) hide show
  1. ai_analyst/ai_analyst.py +1 -1
ai_analyst/ai_analyst.py CHANGED
@@ -11,7 +11,7 @@ from shared.kafka_utils import create_producer, create_consumer
11
  OLLAMA_HOST = os.getenv("OLLAMA_HOST", "") # e.g. http://host.docker.internal:11434
12
  OLLAMA_MODEL = os.getenv("OLLAMA_MODEL", "llama3.1:8b")
13
  HF_TOKEN = os.getenv("HF_TOKEN", "")
14
- HF_MODEL = os.getenv("HF_MODEL", "RayMelius/stockex-analyst")
15
  GROQ_API_KEY = os.getenv("GROQ_API_KEY", "")
16
  GROQ_MODEL = os.getenv("GROQ_MODEL", "llama-3.1-8b-instant")
17
  GROQ_URL = "https://api.groq.com/openai/v1/chat/completions"
 
11
  OLLAMA_HOST = os.getenv("OLLAMA_HOST", "") # e.g. http://host.docker.internal:11434
12
  OLLAMA_MODEL = os.getenv("OLLAMA_MODEL", "llama3.1:8b")
13
  HF_TOKEN = os.getenv("HF_TOKEN", "")
14
+ HF_MODEL = os.getenv("HF_MODEL", "Qwen/Qwen2.5-7B-Instruct")
15
  GROQ_API_KEY = os.getenv("GROQ_API_KEY", "")
16
  GROQ_MODEL = os.getenv("GROQ_MODEL", "llama-3.1-8b-instant")
17
  GROQ_URL = "https://api.groq.com/openai/v1/chat/completions"