File size: 1,912 Bytes
b694b19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
{
  "model_name": "anthropic/claude-opus-4-6",
  "save_sessions": true,
  "session_dataset_repo": "smolagents/ml-intern-sessions",
  "share_traces": true,
  "personal_trace_repo_template": "{hf_user}/ml-intern-sessions",
  "yolo_mode": false,
  "confirm_cpu_jobs": true,
  "auto_file_upload": true,
  "messaging": {
    "enabled": false,
    "auto_event_types": ["approval_required", "error", "turn_complete"],
    "destinations": {}
  },
  "mcpServers": {
    "hf-mcp-server": {
      "transport": "http",
      "url": "https://huggingface.co/mcp?login"
    }
  },
  "_local_provider_examples": {
    "_comment": "Set model_name to any of the prefixes below to use a local / self-hosted LLM. All local providers speak an OpenAI-compatible API. Reasoning effort is automatically disabled for local models.",
    "llamacpp/llama-3-8b": "llama.cpp server (default http://localhost:8080/v1). Override with LLAMACPP_API_BASE env var.",
    "lmstudio/llama-3-8b": "LM Studio (default http://localhost:1234/v1). Override with LMSTUDIO_API_BASE env var.",
    "mlx/llama-3-8b": "MLX OpenAI-compatible wrapper (default http://localhost:8000/v1). Override with MLX_API_BASE env var.",
    "nim/llama-3-8b": "NVIDIA NIM (default http://localhost:8000/v1). Override with NIM_API_BASE env var.",
    "ollama/llama3.1": "Ollama (default http://localhost:11434/v1). Override with OLLAMA_API_BASE env var.",
    "vllm/llama-3-8b": "vLLM (default http://localhost:8000/v1). Override with VLLM_API_BASE env var.",
    "tgi/llama-3-8b": "Text Generation Inference (default http://localhost:8080/v1). Override with TGI_API_BASE env var.",
    "local/llama-3-8b": "Custom PyTorch / any OpenAI-compatible server (default http://localhost:8000/v1). Override with LOCAL_API_BASE env var.",
    "_auth_note": "All local providers use api_key='no-key' by default. Set LOCAL_API_KEY env var if your server requires authentication."
  }
}