queryforge / README.md
Prithvigg's picture
Upload folder using huggingface_hub
039839b verified
metadata
title: QueryForge Environment Server
emoji: πŸ”
colorFrom: blue
colorTo: indigo
sdk: docker
pinned: false
app_port: 8000
base_path: /web
tags:
  - openenv
  - sql
  - reinforcement-learning

QueryForge β€” SQL Debugging & Optimisation Environment

Live environment: https://prithvigg-queryforge.hf.space Interactive demo: https://prithvigg-queryforge.hf.space/demo/

Try it directly below (HF Space viewer only):

SQL is the language that runs the world's data infrastructure. Yet SQL bugs are silent killers β€” a missing JOIN condition inflates totals by 3Γ—, a correlated subquery scans a million rows once per row, a typo in a keyword stops production cold. These bugs are rarely caught by linters, rarely surfaced by error messages, and routinely shipped to production.

QueryForge is an OpenEnv-compatible reinforcement learning environment where an agent learns to debug and optimise SQL queries. The agent receives a broken or slow query, submits fixes, and receives graded feedback from a deterministic DuckDB engine combined with an Anthropic AI quality judge β€” a smooth, informative reward signal across the full 0.0 β†’ 1.0 range.


Why SQL Debugging as an RL Environment?

LLMs can write SQL. What they struggle with is the iterative, feedback-driven debugging loop that real engineers do:

  • Read the error message
  • Form a hypothesis about the root cause
  • Patch the query
  • Check if the output is now correct
  • Refine until it's both correct and efficient

This is precisely the loop that RL is built for. QueryForge provides the environment that closes this loop with a graded, multi-stage reward signal β€” not just "correct / incorrect" but partial credit for syntax validity, execution success, row correctness, and code quality.


Environment Overview

Property Value
Task type SQL debugging & optimisation
Action space Single SQL query string
Observation space Task description + graded feedback
Reward range 0.0 – 1.0 (continuous)
Episode termination Score β‰₯ 0.90, no improvement for 2 steps, or max steps
Grading engine DuckDB (deterministic) + Anthropic AI judge
Concurrent sessions Supported

Reward Scale

The grading pipeline has four stages that produce a smooth partial-progress signal:

Score Meaning
0.00 Syntax error β€” query could not be parsed
0.15 Syntax valid but runtime error
0.30 Executes but returns 0 rows or wrong row count
0.30 – 0.80 Partial row correctness (deterministic, DuckDB)
0.80 – 1.00 Correct rows + AI quality assessment (Anthropic)

The AI judge scores on three axes: Correctness (0–0.50), Optimization (0–0.30 β€” penalises cartesian products, correlated subqueries), Code quality (0–0.20 β€” readability, aliases, formatting).

Offline mode: If ANTHROPIC_API_KEY is not set, the AI judge is skipped and scoring is fully deterministic (capped at 0.80). The done threshold self-adjusts to 0.80 in this case so episodes still terminate correctly.


Action Space

class SQLAction(Action):
    sql: str  # The SQL query to submit for grading

One field. The agent submits a SQL string. No multi-statement queries (; separated) are allowed β€” rejected with score 0.0.


Observation Space

class SQLObservation(Observation):
    # Task context (set on reset, constant within an episode)
    task_id: str            # e.g. "task_easy_syntax"
    task_level: str         # "easy" | "medium" | "hard" | "custom"
    task_title: str         # Human-readable title
    task_description: str   # Full context: schema, broken query, error, goal

    # Per-step grading signals
    syntax_valid: bool      # True if query parsed without error
    execution_success: bool # True if query ran to completion in DuckDB
    execution_error: str    # Runtime error message, if any
    rows_returned: int      # Number of rows returned

    # Feedback
    feedback: str           # Detailed grading feedback (DuckDB + AI judge)
    hint: str               # Actionable hint (suppressed once score >= 0.90)

    # Episode progress
    attempt: int            # Number of queries submitted this episode
    best_score: float       # Highest score achieved so far
    done: bool
    reward: float           # Score for this specific step (0.0 – 1.0)

Built-in Tasks

ID Level Title Max Steps
task_easy_syntax easy Fix Syntax Errors 5
task_medium_join medium Fix the Cartesian JOIN 5
task_hard_cte hard Rewrite Correlated Subquery as CTE 6
task_expert_rank expert Fix the Tie-Breaking Window Function 6
task_expert_recursive expert Traverse Org Chart with Recursive CTE 7
task_expert_window expert Fix Two Broken Window Functions 6

Easy β€” Fix Syntax Errors

Three SQL keywords are misspelled (SELEC, FORM, WEHRE). The agent must identify and correct them.

Schema: users(id, name, age, city) β€” 6 rows Goal: Return name and age of users older than 30 in New York, ordered by name

Medium β€” Fix the Cartesian JOIN

A missing JOIN condition (o.product_id = p.id) causes a cartesian product, inflating every total by 3Γ—. The agent must rewrite using explicit INNER JOIN … ON syntax.

Schema: orders, users, products β€” e-commerce dataset Goal: Correct per-(user, product) total amount spent, ordered by total DESC

Hard β€” Rewrite Correlated Subquery as CTE

A semantically correct but O(NΒ²) query re-executes AVG(salary) for every employee row. The agent must rewrite using a WITH clause that computes department averages exactly once.

Schema: departments, employees β€” 9 employees across 3 departments Goal: Employees who earn strictly above their department average, ordered by dept/salary

Expert β€” Fix the Tie-Breaking Window Function (2 bugs)

Two layered bugs: ROW_NUMBER() drops tied reps AND ORDER BY revenue ASC picks the lowest earners instead of the highest. Agent must fix the sort order AND switch to RANK()/DENSE_RANK() β€” fixing only one still produces wrong results.

Schema: sales_reps(id, name, region, revenue) β€” 6 reps across 2 regions with ties Goal: All reps whose revenue is the highest in their region

Expert β€” Traverse Org Chart with Recursive CTE (2 bugs)

Two layered bugs: the anchor uses WHERE id = 3 (includes VP Eng himself in results) AND the query is a hardcoded two-level CTE that misses deeper employees. Agent must fix the anchor to WHERE manager_id = 3 AND convert to WITH RECURSIVE.

Schema: employees(id, name, manager_id) β€” 14 employees, 4 levels deep Goal: All 8 subordinates of VP Eng at any depth (excluding VP Eng), ordered by id

Expert β€” Fix Broken Window Functions (3 bugs)

Three layered bugs: both SUM and RANK window functions are missing PARTITION BY, they need different ORDER BY clauses, AND the data contains tied revenue values (West Q3=Q4=16000) that must be ranked correctly.

Schema: quarterly_sales(region, quarter, revenue) β€” 8 rows across 2 regions with ties Goal: Per-region running total (ORDER BY quarter) and within-region revenue rank (ORDER BY revenue DESC) with correct tie handling

Structural penalties are enforced per task level/id to prevent gaming:

  • hard: requires WITH clause (βˆ’0.30 if absent)
  • medium: requires explicit JOIN (βˆ’0.20 if absent)
  • task_expert_recursive: requires WITH RECURSIVE (βˆ’0.30) + correct anchor via manager_id (βˆ’0.15)
  • task_expert_rank: penalises ROW_NUMBER() (βˆ’0.20) + penalises ASC ordering without DESC (βˆ’0.15)
  • task_expert_window: requires PARTITION BY in both window functions (βˆ’0.20 if absent, βˆ’0.10 if only one)

Custom Tasks

Register any SQL task at runtime β€” no code changes needed.

Via Python

from tasks import REGISTRY, task_from_dict

REGISTRY.register(task_from_dict({
    "id": "my_window_task",
    "level": "hard",
    "title": "Rank Employees by Salary",
    "schema_ddl": "CREATE TABLE emp (id INT, name VARCHAR, dept VARCHAR, salary DECIMAL); INSERT INTO emp VALUES ...",
    "broken_query": "SELECT name, salary FROM emp ORDER BY salary DESC",
    "expected_rows": [{"name": "Alice", "rank": 1}, ...],
    "hint": "Use ROW_NUMBER() OVER (PARTITION BY dept ORDER BY salary DESC)",
    "solution_query": "SELECT name, RANK() OVER (ORDER BY salary DESC) AS rank FROM emp",
}))

Via REST API (when server is running)

# Register a custom task
curl -X POST http://localhost:8000/tasks \
  -H "Content-Type: application/json" \
  -d '{"id": "my_task", "schema_ddl": "...", "expected_rows": [...]}'

# List all tasks
curl http://localhost:8000/tasks

# Remove a custom task
curl -X DELETE http://localhost:8000/tasks/my_task

Via JSON file

REGISTRY.load_from_json("my_tasks.json")

Quickstart

Install dependencies

python -m venv .venv
.venv/bin/pip install -e ".[dev]"

Run the local playbook (no server needed)

Tests all three built-in tasks directly, with progressive SQL attempts:

ANTHROPIC_API_KEY=your_key .venv/bin/python playbook.py

Run the inference script

Runs any OpenAI-compatible LLM as an agent against all 6 tasks and reports scores:

# Against HuggingFace router
export API_BASE_URL=https://router.huggingface.co/v1
export MODEL_NAME=meta-llama/Llama-3.1-8B-Instruct
export HF_TOKEN=your_hf_token
export ENV_URL=http://127.0.0.1:8000       # or the live HF Space URL
python inference.py

# Against the live HF Space (no local server needed)
export ENV_URL=https://prithvigg-queryforge.hf.space
python inference.py

Run the HTTP server

uvicorn server.app:app --reload --host 0.0.0.0 --port 8000

Baseline Results

The following scores were produced by running meta-llama/Llama-3.1-8B-Instruct (via HuggingFace router) as the agent against all 6 tasks with the full AI judge active.

Task Level Steps Used Best Score
Fix the Syntax Errors easy 1 1.000
Fix the Cartesian JOIN medium 1 0.900
Rewrite Correlated Subquery as CTE hard 1 0.900
Fix the Tie-Breaking Window Function expert 1 1.000
Traverse Org Chart with Recursive CTE expert 2 0.900
Fix Two Broken Window Functions expert 3 0.900
Average 0.933

The easy–hard tasks and the rank/recursive expert tasks were solved in 1–2 steps. The dual-window expert task required 3 steps, demonstrating the feedback loop produces training-relevant multi-step trajectories for harder tasks.


API Endpoints

Method Endpoint Description
POST /reset Start a new episode. Pass {"task_id": "..."} to pin to a task
POST /step Submit a SQL query: {"sql": "SELECT ..."}
GET /state Current episode ID and step count
GET /schema Action and observation JSON schemas
POST /tasks Register a custom task
GET /tasks List all registered tasks
DELETE /tasks/{task_id} Remove a custom task (built-ins protected)
WS /ws WebSocket endpoint for persistent low-latency sessions
GET /health Container health check
GET /docs Interactive OpenAPI documentation

Examples

BASE=https://prithvigg-queryforge.hf.space   # or http://localhost:8000 for local

# Start an episode pinned to the hard task
curl -X POST $BASE/reset \
  -H "Content-Type: application/json" \
  -d '{"task_id": "task_hard_cte"}'

# Submit a query
curl -X POST $BASE/step \
  -H "Content-Type: application/json" \
  -d '{"sql": "WITH dept_avg AS (SELECT department_id, AVG(salary) AS avg_salary FROM employees GROUP BY department_id) SELECT e.name, e.department_id, e.salary FROM employees e JOIN dept_avg d ON e.department_id = d.department_id WHERE e.salary > d.avg_salary ORDER BY e.department_id, e.salary DESC"}'

# List all available tasks
curl $BASE/tasks

Python Client

from queryforge import QueryforgeEnv, SQLAction

with QueryforgeEnv(base_url="http://localhost:8000") as env:
    # Pin to a specific task
    obs = env.reset(task_id="task_medium_join")
    print(obs.task_description)

    # Submit a fix
    result = env.step(SQLAction(sql="""
        SELECT u.name, p.title, SUM(o.amount) AS total_spent
        FROM orders o
        INNER JOIN users u ON o.user_id = u.id
        INNER JOIN products p ON o.product_id = p.id
        GROUP BY u.name, p.title
        ORDER BY total_spent DESC
    """))
    print(f"Score: {result.reward:.3f}")
    print(f"Feedback: {result.observation.feedback}")
    print(f"Done: {result.done}")

    # Register and use a custom task
    env.register_task(TaskSpec(
        id="my_task",
        schema_ddl="CREATE TABLE ...; INSERT INTO ...",
        expected_rows=[{"col": "val"}],
        title="My Custom Task",
    ))
    obs = env.reset(task_id="my_task")

Project Structure

queryforge/
β”œβ”€β”€ __init__.py                     # Public exports (SQLAction, SQLObservation, TaskSpec, REGISTRY)
β”œβ”€β”€ models.py                       # SQLAction, SQLObservation, TaskSpec Pydantic models
β”œβ”€β”€ tasks.py                        # Built-in tasks (easyβ†’expert) + thread-safe TaskRegistry
β”œβ”€β”€ judge.py                        # 4-stage grading pipeline (DuckDB + Anthropic)
β”œβ”€β”€ client.py                       # QueryforgeEnv client with task management helpers
β”œβ”€β”€ playbook.py                     # Local test runner (no server required)
β”œβ”€β”€ inference.py                    # Baseline inference script (any OpenAI-compatible LLM)
β”œβ”€β”€ demo.py                         # Gradio interactive demo (mounted at /demo)
β”œβ”€β”€ Dockerfile                      # Container image
β”œβ”€β”€ openenv.yaml                    # OpenEnv manifest
β”œβ”€β”€ pyproject.toml                  # Project metadata and dependencies
β”œβ”€β”€ uv.lock                         # Locked dependencies
└── server/
    β”œβ”€β”€ app.py                      # FastAPI app β€” core + /tasks REST endpoints + Gradio mount
    β”œβ”€β”€ queryforge_environment.py   # Environment class (reset, step, state)
    └── requirements.txt            # Server dependencies

Deployment

Hugging Face Spaces (recommended)

UV_CACHE_DIR=/tmp/uv-cache openenv push . --repo-id <hf-username>/queryforge

Add ANTHROPIC_API_KEY as a Space secret after deployment. Without it, the environment runs in deterministic-only mode (scores capped at 0.80, done threshold self-adjusts accordingly).

Docker

docker build -t queryforge:latest .
docker run -p 8000:8000 -e ANTHROPIC_API_KEY=$ANTHROPIC_API_KEY queryforge:latest

The deployed environment exposes:

  • /web β€” Interactive UI for exploring the environment
  • /docs β€” Full OpenAPI / Swagger interface
  • /ws β€” WebSocket endpoint for persistent agent sessions
  • /health β€” Container health monitoring

Environment Design Notes

Why DuckDB? DuckDB runs fully in-memory with no external process or network dependency. Each step() call creates an isolated connection, seeds it with the task's schema, runs the agent's query, then closes β€” complete isolation with zero shared state between steps.

Why a 4-stage reward? Binary correct/incorrect rewards give an agent no gradient to climb when its query is simply broken. The 4-stage pipeline means every improvement β€” fixing a typo, avoiding a runtime error, returning the right row count, getting the right rows, writing clean SQL β€” is rewarded. This produces a smooth loss landscape for policy gradient methods.

Why structural penalties? Without them, an agent could achieve 0.80 on the hard CTE task by submitting the original correlated subquery verbatim (rows match, but the task was never solved). Structural penalties enforce that the agent actually learned what to change, not just that rows matched.