# ARIA Project Status _Last updated: May 5, 2026_ ## What Was Built Multi-agent pipeline in agent/agent.py with 5 stages: 1. Query Architect: generates 5 MeSH-optimised PubMed queries via Qwen2.5-72B on AMD MI300X 2. Literature Scout: fetches from PubMed and Europe PMC in parallel, deduplicates by PMID 3. PRISMA Filter: automatic inclusion/exclusion screening with one-line reasons, user can override any decision 4. Evidence Synthesiser: structured synthesis with Background, Key Findings, Level of Evidence, Conflicting Evidence, Research Gaps, Clinical Implications 5. Citation Builder: formatted references with PMID, synthesis runs on PRISMA-included papers only ## Additional Features * SSE streaming: real-time 5-stage progress bar with percentage * PRISMA-style paper screening: automated include/exclude with rationale, user override buttons * Evidence confidence scoring: green/yellow/red badges on each synthesis section with hover tooltips * Abstract viewer: click any citation to expand full abstract inline * PDF export: download formatted report via ReportLab * Selective literature review: checkboxes on citations, user picks papers, generates focused academic paragraph * Predictive model: constructive and destructive forecasts as a post-synthesis stage * Evidence comparison table: LLM extracts structured data table with real author names, no duplicates * Follow-up Q&A: ask follow-up questions after synthesis, answered using already-fetched papers * Query refinement suggestions: 3 AI-generated follow-up research questions based on synthesis gaps * Session history: queries saved to sessions.json, reloadable from sidebar * Rate limit retry logic: automatic backoff on API errors * SSL patch for PubMed and Europe PMC Entrez on corporate/university networks ## Tech Stack * LLM: Qwen2.5-72B-Instruct on AMD MI300X via vLLM 0.17.1 * Agent Framework: LangGraph + LangChain * Literature Retrieval: BioPython Entrez / PubMed NCBI + Europe PMC * Web Framework: Flask with SSE streaming * PDF: ReportLab * Frontend: HTML, CSS, vanilla JS * Runtime: Windows 11, Python 3.11, RTX 3060 12GB (local) + AMD MI300X 192GB (inference) ## API Endpoints * GET / — main UI * GET /stream — SSE pipeline stream * POST /query — standard pipeline (fallback) * POST /score — confidence scoring * POST /predict — predictive model * POST /selective-review — focused literature review from selected papers * POST /extract-table — evidence comparison table * POST /export-pdf — PDF report download * POST /followup — follow-up question against existing synthesis * POST /suggest-queries — 3 AI-generated follow-up research questions * GET /sessions — load query history * POST /sessions/save — save completed query ## Environment * VLLM_BASE_URL: set via HF Space environment variable * VLLM_API_KEY: EMPTY * Python venv at ./venv * Start server: venv\Scripts\activate && python app.py * AMD MI300X: DigitalOcean droplet, 192GB VRAM ## Deployment * HF Space: https://huggingface.co/spaces/lablab-ai-amd-developer-hackathon/glitch-squad-biomedical-assistant * GitHub: https://github.com/azlaan428/glitch-squad-biomedical-assistant ## What Remains 1. Demo video (under 5 minutes) 2. Final lablab.ai submission