# Experiment 1 exploratory Cell D — Agent-as-Tool + optimized MCP + # model-side serving optimizations. # # Cell C keeps model precision fixed and measures recoverable MCP transport # overhead. Cell D intentionally changes model-side serving too: # - optimized AaT MCP batch runner + connection reuse # - vLLM prefix caching # - compressed-tensors W8A8 INT8 checkpoint # - BF16 model dtype, which unlocks fp8 KV cache on vLLM 0.19.0 # # This is a follow-on / "best-effort optimized serving" condition, not part of # the clean A/B/C transport-only headline. EXPERIMENT_NAME="aat_mcp_model_optimized" EXPERIMENT_CELL="D" EXPERIMENT_FAMILY="exp1_model_optimization" SCENARIO_SET_NAME="smartgrid_multi_domain" SCENARIOS_GLOB="data/scenarios/multi_*.json" MODEL_ID="openai/Llama-3.1-8B-Instruct-int8" ORCHESTRATION="agent_as_tool" MCP_MODE="optimized" TRIALS=3 ENABLE_SMARTGRID_SERVERS=1 CONTRIBUTING_EXPERIMENTS="exp1_model_optimization" SCENARIO_DOMAIN_SCOPE="multi_domain" MODEL_PROVIDER="vllm" SERVING_STACK="insomnia_vllm" QUANTIZATION_MODE="compressed-tensors-int8-bf16-fp8kv" # Keep the first exploratory D run aligned with the proven INT8 smoke path. # If this hits context length under full AaT prompts, rerun D at 16384/32768. MAX_MODEL_LEN=8192 TEMPERATURE=0.0 MAX_TOKENS=0 LAUNCH_VLLM=1 VLLM_MODEL_PATH="models/Llama-3.1-8B-Instruct-int8" VLLM_SERVED_MODEL_NAME="Llama-3.1-8B-Instruct-int8" VLLM_DTYPE="bfloat16" VLLM_PORT=8000 VLLM_ENABLE_AUTO_TOOL_CHOICE=1 VLLM_TOOL_CALL_PARSER="llama3_json" VLLM_STARTUP_TIMEOUT=1200 # Model-side optimization knobs (#29 + #30 follow-on): # --quantization compressed-tensors: RedHatAI W8A8 checkpoint, smoke-proven # in Slurm job 8979660. # --kv-cache-dtype fp8: compatible with the BF16 dtype used by the INT8 path. # --enable-prefix-caching: same prompt-prefix reuse used by Cell C. EXTRA_VLLM_ARGS="--quantization compressed-tensors --kv-cache-dtype fp8 --enable-prefix-caching" ENABLE_WANDB=1 WANDB_ENTITY="assetopsbench-smartgrid" WANDB_PROJECT="assetopsbench-smartgrid" WANDB_MODE="online" AAT_MCP_SERVER_LAUNCH_MODE="python" AAT_MCP_CLIENT_TIMEOUT_SECONDS=120 AAT_PARALLEL_TOOL_CALLS=false # Torch profiler: captures one replay pass per run while vLLM is still live. TORCH_PROFILE="${TORCH_PROFILE:-1}"