OPENENV_RL_01 / .env.example
Siddharaj Shirke
deploy: fresh snapshot to Hugging Face Space
3eae4cc
# Gov Workflow OpenEnv
# Detailed environment template for local run, E2E validation, Docker preflight,
# and release deployment.
#
# Usage:
# 1) Copy this file to .env
# 2) Fill only the auth/provider values you use
# 3) Keep defaults unless you intentionally need different behavior
# -----------------------------------------------------------------------------
# 1) LLM Provider Endpoints and Auth
# -----------------------------------------------------------------------------
# Primary OpenAI-compatible endpoint used by inference/simulation runtime.
API_BASE_URL=https://integrate.api.nvidia.com/v1
# OpenAI-compatible model used for LLM inference mode.
MODEL_NAME=meta/llama-3.3-70b-instruct
# Auth precedence in runtime:
# HF_TOKEN -> OPENAI_API_KEY -> API_KEY
HF_TOKEN=
OPENAI_API_KEY=
API_KEY=
# Optional image tag used by inference / utility flows.
LOCAL_IMAGE_NAME=gov-workflow-openenv:latest
# Inference acceptance criteria (inference.py).
MAX_STEPS=80
SUCCESS_SCORE_THRESHOLD=0.50
# -----------------------------------------------------------------------------
# 2) Provider-Specific API Base URLs
# -----------------------------------------------------------------------------
# OpenAI-compatible provider URL (fallback path in engine/simulator).
OPENAI_API_BASE_URL=https://api.openai.com/v1
# NVIDIA provider URL for NIM calls.
NVIDIA_API_BASE_URL=https://integrate.api.nvidia.com/v1
# -----------------------------------------------------------------------------
# 3) Model Routing and Fallback Pools
# -----------------------------------------------------------------------------
# Optional CSV fallback models for OpenAI-compatible runtime.
# Example:
# MODEL_FALLBACKS=meta/llama-3.1-8b-instruct,microsoft/phi-4-mini-instruct
MODEL_FALLBACKS=
# Optional CSV fallback models for NVIDIA runtime.
NVIDIA_MODEL_FALLBACKS=
# Primary NVIDIA model for NVIDIA-key runtime path.
NVIDIA_MODEL=meta/llama-3.3-70b-instruct
# NVIDIA keys for baseline and simulation fallback behavior.
# Get keys at: https://build.nvidia.com/explore/discover
NVIDIA_API_KEY=
NVIDIA_API_KEY_2=
# -----------------------------------------------------------------------------
# 4) Environment Transport (Direct vs HTTP)
# -----------------------------------------------------------------------------
# Used by inference / gateway code.
# Allowed: auto, http, direct
OPENENV_ENV_TRANSPORT=auto
# Base URL for HTTP transport path.
OPENENV_ENV_BASE_URL=http://127.0.0.1:7860
# Optional explicit API prefix for /reset /step /grade calls.
# Typical values: (empty), /api, /api/v1
OPENENV_ENV_API_PREFIX=
# Optional candidate prefixes (CSV) tried before built-in fallback sequence.
# Example: /api/v1,/api
OPENENV_ENV_API_PREFIX_CANDIDATES=
# Force HTTP/FastAPI gateway even when direct transport is available.
# Allowed truthy values: 1, true, yes, on
FORCE_FASTAPI_GATEWAY=0
# -----------------------------------------------------------------------------
# 5) Structured API Alias Controls (app.main)
# -----------------------------------------------------------------------------
# Enables automatic aliasing from source prefix to versioned prefix.
ENABLE_STRUCTURED_V1_API=1
OPENENV_API_SOURCE_PREFIX=/api
OPENENV_API_V1_PREFIX=/api/v1
# -----------------------------------------------------------------------------
# 6) FastAPI Server Settings (SERVER_* in app/config.py)
# -----------------------------------------------------------------------------
SERVER_HOST=0.0.0.0
SERVER_PORT=7860
SERVER_LOG_LEVEL=info
# Keep 1 for in-memory session store unless external shared state is added.
SERVER_WORKERS=1
# JSON list string expected by Pydantic settings.
SERVER_CORS_ORIGINS=["*"]
# -----------------------------------------------------------------------------
# 7) Environment Defaults (ENV_* in app/config.py)
# -----------------------------------------------------------------------------
ENV_DEFAULT_TASK_ID=district_backlog_easy
ENV_DEFAULT_SEED=11
ENV_MAX_SESSIONS=100
ENV_MAX_STEPS_PER_EPISODE=500
# -----------------------------------------------------------------------------
# 8) Runtime Throttling
# -----------------------------------------------------------------------------
# Delay between LLM calls used by baseline_openai.py.
LLM_CALL_DELAY=12.0
# -----------------------------------------------------------------------------
# 9) Persistence and Storage
# -----------------------------------------------------------------------------
# Enables SQLite/filesystem persistence.
STORAGE_ENABLED=true
# Preferred persistence root (used by app/persistence.py).
# Local example: C:/Users/your-user/OPENENV_RL/outputs/persist
# HF Spaces example: /data/openenv_rl
OPENENV_DATA_DIR=
# Legacy fallback path key still supported by code.
STORAGE_DATA_DIR=
# -----------------------------------------------------------------------------
# 10) Frontend Dev Proxy (Vite)
# -----------------------------------------------------------------------------
# Used by frontend/react/vite.config.js for local /api proxy target.
VITE_DEV_API_TARGET=http://127.0.0.1:7860