Upload configs/cli_agent_config.json
Browse files
configs/cli_agent_config.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_name": "anthropic/claude-opus-4-6",
|
| 3 |
+
"save_sessions": true,
|
| 4 |
+
"session_dataset_repo": "smolagents/ml-intern-sessions",
|
| 5 |
+
"share_traces": true,
|
| 6 |
+
"personal_trace_repo_template": "{hf_user}/ml-intern-sessions",
|
| 7 |
+
"yolo_mode": false,
|
| 8 |
+
"confirm_cpu_jobs": true,
|
| 9 |
+
"auto_file_upload": true,
|
| 10 |
+
"messaging": {
|
| 11 |
+
"enabled": false,
|
| 12 |
+
"auto_event_types": ["approval_required", "error", "turn_complete"],
|
| 13 |
+
"destinations": {}
|
| 14 |
+
},
|
| 15 |
+
"mcpServers": {
|
| 16 |
+
"hf-mcp-server": {
|
| 17 |
+
"transport": "http",
|
| 18 |
+
"url": "https://huggingface.co/mcp?login"
|
| 19 |
+
}
|
| 20 |
+
},
|
| 21 |
+
"_local_provider_examples": {
|
| 22 |
+
"_comment": "Set model_name to any of the prefixes below to use a local / self-hosted LLM. All local providers speak an OpenAI-compatible API. Reasoning effort is automatically disabled for local models.",
|
| 23 |
+
"llamacpp/llama-3-8b": "llama.cpp server (default http://localhost:8080/v1). Override with LLAMACPP_API_BASE env var.",
|
| 24 |
+
"lmstudio/llama-3-8b": "LM Studio (default http://localhost:1234/v1). Override with LMSTUDIO_API_BASE env var.",
|
| 25 |
+
"mlx/llama-3-8b": "MLX OpenAI-compatible wrapper (default http://localhost:8000/v1). Override with MLX_API_BASE env var.",
|
| 26 |
+
"nim/llama-3-8b": "NVIDIA NIM (default http://localhost:8000/v1). Override with NIM_API_BASE env var.",
|
| 27 |
+
"ollama/llama3.1": "Ollama (default http://localhost:11434/v1). Override with OLLAMA_API_BASE env var.",
|
| 28 |
+
"vllm/llama-3-8b": "vLLM (default http://localhost:8000/v1). Override with VLLM_API_BASE env var.",
|
| 29 |
+
"tgi/llama-3-8b": "Text Generation Inference (default http://localhost:8080/v1). Override with TGI_API_BASE env var.",
|
| 30 |
+
"local/llama-3-8b": "Custom PyTorch / any OpenAI-compatible server (default http://localhost:8000/v1). Override with LOCAL_API_BASE env var.",
|
| 31 |
+
"_auth_note": "All local providers use api_key='no-key' by default. Set LOCAL_API_KEY env var if your server requires authentication."
|
| 32 |
+
}
|
| 33 |
+
}
|