Spaces:
Sleeping
Sleeping
| title: Deep Research Agent | |
| emoji: 𧬠| |
| colorFrom: red | |
| colorTo: yellow | |
| sdk: gradio | |
| app_file: app.py | |
| pinned: false | |
| # Deep Research Agents π | |
| An agentic research pipeline that helps you clarify topics, plan research, and automatically gather information from the web to generate comprehensive Markdown reports. | |
| ## π Features | |
| - **Topic Clarification**: Iteratively refines broad research questions into specific, actionable topics. | |
| - **Strategic Planning**: Generates structured research plans to cover all necessary aspects of a topic. | |
| - **Agentic Coordination**: Uses `smolagents` and **Tavily Search** to orchestrate specialized agents that browse the web and synthesize findings. | |
| - **Robust Model Support**: Specifically optimized for "Reasoning" models (like DeepSeek-R1) and stable tool-calling models (like Qwen-2.5-Coder and MiniMax). | |
| - **Corporate Network Ready**: Includes automated SSL certificate handling via `truststore` to bypass common proxy errors. | |
| ## π οΈ Project Structure | |
| ```text | |
| researcher-agents/ | |
| βββ main.py # CLI Entry point | |
| βββ app.py # Streamlit UI | |
| βββ requirements.txt # Dependencies | |
| βββ src/ # Core Agents | |
| βββ clarifier.py # Topic refinement | |
| βββ planner.py # Strategic planning | |
| βββ splitter.py # Task decomposition | |
| βββ coordinator.py # Agent orchestration | |
| βββ prompts.py # LLM Instructions | |
| ``` | |
| ## βοΈ Setup | |
| 1. **Clone the repository**: | |
| ```bash | |
| git clone https://github.com/vizcayal/researcher_agents.git | |
| cd researcher_agents | |
| ``` | |
| 2. **Create a virtual environment**: | |
| ```bash | |
| python -m venv .venv | |
| .\.venv\Scripts\activate | |
| ``` | |
| 3. **Install dependencies**: | |
| ```bash | |
| pip install -r requirements.txt | |
| ``` | |
| 4. **Environment Variables**: | |
| Create a `.env` file in the root directory: | |
| ```env | |
| HF_KEY=your_huggingface_token | |
| FIRECRAWL_KEY=your_firecrawl_api_key | |
| ``` | |
| ## π Usage | |
| ### Command Line Interface | |
| Run the full automated research pipeline: | |
| ```bash | |
| python main.py | |
| ``` | |
| ### Gradio Web App | |
| Run the interactive UI: | |
| ```bash | |
| python app.py | |
| ``` | |
| ## π€ Recommended Models | |
| The project is configured to work with the Hugging Face Serverless Inference API: | |
| - **Reasoning**: `deepseek-ai/DeepSeek-R1-Distill-Llama-8B` | |
| - **Orchestration**: `Qwen/Qwen2.5-Coder-32B-Instruct` | |
| - **Output**: `meta-llama/Llama-3.3-70B-Instruct` | |
| ## π License | |
| MIT | |