| """LLM client instantiation and provider detection.""" |
|
|
| import os |
| import threading |
| from typing import Any, Dict, Optional, Tuple |
|
|
| from llm_interface import ( |
| LLMInterface, |
| GPT, |
| Gemini, |
| Claude, |
| Claude_Opus, |
| Claude_Sonnet_4_5, |
| DeepSeek, |
| Grok, |
| ) |
|
|
| |
| MODEL_CONFIG_SUMMARY: Dict[str, Dict[str, Any]] = {} |
| MODEL_CONFIG_LOCK = threading.Lock() |
|
|
|
|
| def infer_provider_and_model(model: str) -> Tuple[str, str]: |
| """Parse model string into provider hint and actual model name.""" |
| normalized = model.strip() |
| if "/" in normalized: |
| provider, actual_model = normalized.split("/", 1) |
| else: |
| provider, actual_model = "", normalized |
| return provider.lower(), actual_model.strip() |
|
|
|
|
| def detect_provider(model: str, actual_model_lower: Optional[str] = None) -> str: |
| """Detect the provider for a given model string.""" |
| provider_hint, actual_model = infer_provider_and_model(model) |
| actual_lower = actual_model_lower or actual_model.lower() |
|
|
| if (provider_hint in {"", "openai", "azure", "azure_openai"}) and actual_lower.startswith("gpt"): |
| return "openai" |
| if provider_hint in {"gemini", "google"} or "gemini" in actual_lower: |
| return "google" |
| if provider_hint == "anthropic" or "claude" in actual_lower: |
| return "anthropic" |
| if provider_hint == "xai" or "grok" in actual_lower: |
| return "xai" |
| if provider_hint == "deepseek" or "deepseek" in actual_lower: |
| return "deepseek" |
| return provider_hint or "openai" |
|
|
|
|
| def instantiate_llm_client( |
| model: str, |
| *, |
| is_reasoning_model: bool, |
| timeout: float, |
| base_url: Optional[str], |
| api_key: Optional[str], |
| ) -> Tuple[LLMInterface, Dict[str, Any]]: |
| """Create an LLM client instance for the given model.""" |
| provider_hint, actual_model = infer_provider_and_model(model) |
| actual_model_lower = actual_model.lower() |
| provider = detect_provider(model, actual_model_lower) |
| config: Dict[str, Any] = { |
| "requested_model": model, |
| "actual_model": actual_model, |
| "reasoning_mode": is_reasoning_model, |
| } |
|
|
| |
| if provider == "openrouter": |
| requested_lower = model.lower().strip() |
| if requested_lower in {"gemini 3", "gemini3"}: |
| or_slug = "google/gemini-3-pro-preview" |
| elif "/" in model: |
| or_slug = model |
| else: |
| if actual_model_lower.startswith("gemini-3"): |
| or_slug = f"google/{actual_model}" |
| else: |
| or_slug = "google/gemini-3-pro-preview" |
|
|
| openrouter_base = "https://openrouter.ai/api/v1" |
| resolved_key = api_key or os.getenv("OPENROUTER_API_KEY") |
| reasoning_effort = "high" if is_reasoning_model else None |
| client = GPT( |
| model=or_slug, |
| reasoning_effort=reasoning_effort, |
| timeout=timeout, |
| base_url=openrouter_base, |
| api_key=resolved_key, |
| ) |
| config.update({ |
| "provider": "openrouter", |
| "interface": client.__class__.__name__, |
| "reasoning_effort": reasoning_effort, |
| "base_url": openrouter_base, |
| "openrouter_model_slug": or_slug, |
| }) |
|
|
| elif provider == "openai": |
| reasoning_effort = "high" if is_reasoning_model else None |
| client = GPT( |
| model=actual_model, |
| reasoning_effort=reasoning_effort, |
| timeout=timeout, |
| base_url=base_url, |
| api_key=api_key, |
| ) |
| config.update({ |
| "provider": provider, |
| "interface": client.__class__.__name__, |
| "reasoning_effort": reasoning_effort, |
| "base_url": base_url or "https://api.openai.com/v1", |
| }) |
| elif provider == "google": |
| client = Gemini(model=actual_model, timeout=timeout, api_key=api_key) |
| config.update({ |
| "provider": provider, |
| "interface": client.__class__.__name__, |
| "reasoning_effort": None, |
| }) |
| elif provider == "anthropic": |
| if "claude-sonnet-4-5" in actual_model_lower: |
| client = Claude_Sonnet_4_5(model=actual_model, api_key=api_key) |
| elif "claude-opus" in actual_model_lower: |
| client = Claude_Opus(model=actual_model, api_key=api_key) |
| else: |
| client = Claude(model=actual_model, api_key=api_key) |
| config.update({ |
| "provider": provider, |
| "interface": client.__class__.__name__, |
| "reasoning_effort": "thinking-enabled", |
| }) |
| elif provider == "xai": |
| reasoning_effort = "high" if is_reasoning_model else None |
| client = Grok( |
| model=actual_model, |
| reasoning_effort=reasoning_effort, |
| timeout=timeout, |
| api_key=api_key, |
| ) |
| config.update({ |
| "provider": provider, |
| "interface": client.__class__.__name__, |
| "reasoning_effort": reasoning_effort, |
| "base_url": "https://api.x.ai/v1", |
| }) |
| elif provider == "deepseek": |
| client = DeepSeek( |
| model=actual_model, |
| timeout=timeout, |
| api_key=api_key, |
| ) |
| config.update({ |
| "provider": provider, |
| "interface": client.__class__.__name__, |
| "reasoning_effort": None, |
| "base_url": "https://api.deepseek.com", |
| }) |
| else: |
| raise ValueError(f"Unsupported model identifier '{model}' for llm_interface integration.") |
|
|
| if api_key: |
| config["api_key_hint"] = f"***{api_key[-6:]}" |
|
|
| with MODEL_CONFIG_LOCK: |
| MODEL_CONFIG_SUMMARY[model] = config |
|
|
| return client, config |
|
|