# Standard OpenEnv / inference variables API_BASE_URL=https://integrate.api.nvidia.com/v1 MODEL_NAME=meta/llama-3.3-70b-instruct HF_TOKEN= OPENAI_API_KEY= API_KEY= LOCAL_IMAGE_NAME=gov-workflow-openenv:latest MAX_STEPS=80 SUCCESS_SCORE_THRESHOLD=0.50 # Provider-specific API base URLs (used by frontend simulation bridge) OPENAI_API_BASE_URL=https://api.openai.com/v1 NVIDIA_API_BASE_URL=https://integrate.api.nvidia.com/v1 # Optional fallback model lists (comma-separated) MODEL_FALLBACKS= NVIDIA_MODEL_FALLBACKS= # NVIDIA Build API (fallback / internal) # Copy this file to .env and fill in your values # Get your key at: https://build.nvidia.com/explore/discover NVIDIA_API_KEY=nvapi-your-key-here NVIDIA_API_KEY_2= # LLM Model Selection NVIDIA_MODEL=meta/llama-3.3-70b-instruct # Server Settings SERVER_HOST=0.0.0.0 SERVER_PORT=7860 SERVER_LOG_LEVEL=info SERVER_WORKERS=1 # Environment Settings ENV_DEFAULT_TASK_ID=district_backlog_easy ENV_DEFAULT_SEED=11 ENV_MAX_SESSIONS=100 ENV_MAX_STEPS_PER_EPISODE=500 # API Throttling LLM_CALL_DELAY=12.0 # Persistence (SQLite + filesystem) # For Hugging Face persistent storage, set OPENENV_DATA_DIR=/data/openenv_rl STORAGE_ENABLED=true OPENENV_DATA_DIR= # Optional CSV list of extra model directories to scan for RL checkpoints. # Example: # OPENENV_MODEL_SEARCH_DIRS=/data/openenv_rl/results/best_model/phase1,/data/openenv_rl/results/best_model/phase2 OPENENV_MODEL_SEARCH_DIRS=