ARIA Project Status
Last updated: May 5, 2026
What Was Built
Multi-agent pipeline in agent/agent.py with 5 stages:
- Query Architect: generates 5 MeSH-optimised PubMed queries via Qwen2.5-72B on AMD MI300X
- Literature Scout: fetches from PubMed and Europe PMC in parallel, deduplicates by PMID
- PRISMA Filter: automatic inclusion/exclusion screening with one-line reasons, user can override any decision
- Evidence Synthesiser: structured synthesis with Background, Key Findings, Level of Evidence, Conflicting Evidence, Research Gaps, Clinical Implications
- Citation Builder: formatted references with PMID, synthesis runs on PRISMA-included papers only
Additional Features
- SSE streaming: real-time 5-stage progress bar with percentage
- PRISMA-style paper screening: automated include/exclude with rationale, user override buttons
- Evidence confidence scoring: green/yellow/red badges on each synthesis section with hover tooltips
- Abstract viewer: click any citation to expand full abstract inline
- PDF export: download formatted report via ReportLab
- Selective literature review: checkboxes on citations, user picks papers, generates focused academic paragraph
- Predictive model: constructive and destructive forecasts as a post-synthesis stage
- Evidence comparison table: LLM extracts structured data table with real author names, no duplicates
- Follow-up Q&A: ask follow-up questions after synthesis, answered using already-fetched papers
- Query refinement suggestions: 3 AI-generated follow-up research questions based on synthesis gaps
- Session history: queries saved to sessions.json, reloadable from sidebar
- Rate limit retry logic: automatic backoff on API errors
- SSL patch for PubMed and Europe PMC Entrez on corporate/university networks
Tech Stack
- LLM: Qwen2.5-72B-Instruct on AMD MI300X via vLLM 0.17.1
- Agent Framework: LangGraph + LangChain
- Literature Retrieval: BioPython Entrez / PubMed NCBI + Europe PMC
- Web Framework: Flask with SSE streaming
- PDF: ReportLab
- Frontend: HTML, CSS, vanilla JS
- Runtime: Windows 11, Python 3.11, RTX 3060 12GB (local) + AMD MI300X 192GB (inference)
API Endpoints
- GET / � main UI
- GET /stream � SSE pipeline stream
- POST /query � standard pipeline (fallback)
- POST /score � confidence scoring
- POST /predict � predictive model
- POST /selective-review � focused literature review from selected papers
- POST /extract-table � evidence comparison table
- POST /export-pdf � PDF report download
- POST /followup � follow-up question against existing synthesis
- POST /suggest-queries � 3 AI-generated follow-up research questions
- GET /sessions � load query history
- POST /sessions/save � save completed query
Environment
- VLLM_BASE_URL: set via HF Space environment variable
- VLLM_API_KEY: EMPTY
- Python venv at ./venv
- Start server: venv\Scripts\activate && python app.py
- AMD MI300X: DigitalOcean droplet, 192GB VRAM
Deployment
- HF Space: https://huggingface.co/spaces/lablab-ai-amd-developer-hackathon/glitch-squad-biomedical-assistant
- GitHub: https://github.com/azlaan428/glitch-squad-biomedical-assistant
What Remains
- Demo video (under 5 minutes)
- Final lablab.ai submission