Spaces:
Running
Running
| title: Automatic Post — Agents | |
| emoji: ✍️ | |
| colorFrom: purple | |
| colorTo: blue | |
| sdk: docker | |
| app_port: 7860 | |
| pinned: false | |
| # Agents service (CrewAI + LiteLLM + Hugging Face) | |
| FastAPI app that exposes: | |
| - `GET /health` — liveness check | |
| - `POST /generate` — body: `topic`, optional `feedback`, `memory_context`, `tone_instruction`; returns `{"post": "..."}` | |
| Set **`HF_TOKEN`** (or **`HUGGINGFACE_HUB_TOKEN`**) as a [Space secret](https://huggingface.co/docs/hub/spaces-overview#managing-secrets) (required). Set **`AGENTS_LLM_MODEL`** as a Space **variable** (required — e.g. `huggingface/Qwen/Qwen2.5-7B-Instruct`; see `.env.example`). Optionally set **`AGENTS_LLM_TEMPERATURE`** or **`AGENTS_LLM_BASE_URL`** for an OpenAI-compatible endpoint (see `crew.py`). Do **not** use WebWorld with the public HF router only — it is not supported there unless **`AGENTS_LLM_BASE_URL`** points at your own endpoint. | |
| **Port:** the container listens on `$PORT` when the platform sets it, otherwise **7860** (Hugging Face Spaces default). | |
| For local dev without Docker you can still run `uvicorn service:app --port 9000` if you prefer. | |