| import os |
| from typing import Any, Callable |
|
|
| from smolagents import HfApiModel, InferenceClientModel, LiteLLMModel, OpenAIServerModel |
|
|
|
|
| def get_huggingface_api_model(model_id: str, **kwargs) -> HfApiModel: |
| """ |
| Returns a Hugging Face API model instance. |
| Args: |
| model_id (str): The model identifier. |
| **kwargs: Additional keyword arguments for the model. |
| Returns: |
| HfApiModel: Hugging Face API model instance. |
| """ |
| api_key = os.getenv("HUGGINGFACEHUB_API_TOKEN") |
| if not api_key: |
| raise ValueError("HUGGINGFACEHUB_API_TOKEN is not set") |
|
|
| return HfApiModel(model_id=model_id, token=api_key, **kwargs) |
|
|
|
|
| def get_inference_client_model(model_id: str, **kwargs) -> InferenceClientModel: |
| """ |
| Returns an Inference Client model instance. |
| Args: |
| model_id (str): The model identifier. |
| **kwargs: Additional keyword arguments for the model. |
| Returns: |
| InferenceClientModel: Inference client model instance. |
| """ |
| api_key = os.getenv("HUGGINGFACEHUB_API_TOKEN") |
| if not api_key: |
| raise ValueError("HUGGINGFACEHUB_API_TOKEN is not set") |
|
|
| return InferenceClientModel(model_id=model_id, token=api_key, **kwargs) |
|
|
|
|
| def get_openai_server_model(model_id: str, **kwargs) -> OpenAIServerModel: |
| """ |
| Returns an OpenAI server model instance. |
| Args: |
| model_id (str): The model identifier. |
| **kwargs: Additional keyword arguments for the model. |
| Returns: |
| OpenAIServerModel: OpenAI server model instance. |
| """ |
| api_key = os.getenv("OPENAI_API_KEY") |
| if not api_key: |
| raise ValueError("OPENAI_API_KEY is not set") |
|
|
| |
| |
| |
|
|
| return OpenAIServerModel( |
| model_id=model_id, api_key=api_key, api_base=api_base, **kwargs |
| ) |
|
|
|
|
| def get_lite_llm_model(model_id: str, **kwargs) -> LiteLLMModel: |
| """ |
| Returns a LiteLLM model instance. |
| Args: |
| model_id (str): The model identifier. |
| **kwargs: Additional keyword arguments for the model. |
| Returns: |
| LiteLLMModel: LiteLLM model instance. |
| """ |
| return LiteLLMModel(model_id=model_id, **kwargs) |
|
|
|
|
| def get_model(model_type: str, model_id: str, **kwargs) -> Any: |
| """ |
| Returns a model instance based on the specified type. |
| Args: |
| model_type (str): The type of the model (e.g., 'HfApiModel'). |
| model_id (str): The model identifier. |
| **kwargs: Additional keyword arguments for the model. |
| Returns: |
| Any: Model instance of the specified type. |
| """ |
| models: dict[str, Callable[..., Any]] = { |
| "HfApiModel": get_huggingface_api_model, |
| "InferenceClientModel": get_inference_client_model, |
| "OpenAIServerModel": get_openai_server_model, |
| "LiteLLMModel": get_lite_llm_model, |
| } |
|
|
| if model_type not in models: |
| raise ValueError(f"Unknown model type: {model_type}") |
|
|
| return models[model_type](model_id, **kwargs) |