| """Unit tests for scripts/aat_runner.py. |
| |
| These tests use stubbed RunResult objects and mocked Runner.run() so |
| they don't require network, MCP subprocesses, or WatsonX. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import argparse |
| import asyncio |
| import json |
| import logging |
| import sys |
| from pathlib import Path |
| from types import SimpleNamespace |
| from unittest.mock import patch |
|
|
| import pytest |
|
|
|
|
| def _stub_args(**overrides): |
| defaults = dict( |
| prompt="test prompt", |
| output="/tmp/out.json", |
| model_id="watsonx/meta-llama/llama-3-3-70b-instruct", |
| mcp_mode="direct", |
| max_turns=30, |
| parallel_tool_calls=False, |
| verbose=False, |
| ) |
| defaults.update(overrides) |
| return argparse.Namespace(**defaults) |
|
|
|
|
| def _msg_item(text: str): |
| """Build a stub message_output_item matching AOB's parser.""" |
| return SimpleNamespace( |
| type="message_output_item", |
| raw_item=SimpleNamespace(content=[SimpleNamespace(text=text)]), |
| ) |
|
|
|
|
| def _tool_call_item(name: str, args: dict, call_id: str = "c1"): |
| """Build a stub tool_call_item.""" |
| import json as _json |
|
|
| return SimpleNamespace( |
| type="tool_call_item", |
| raw_item=SimpleNamespace( |
| name=name, |
| arguments=_json.dumps(args), |
| call_id=call_id, |
| id=call_id, |
| ), |
| ) |
|
|
|
|
| def _tool_output_item(output: str): |
| return SimpleNamespace(type="tool_call_output_item", output=output) |
|
|
|
|
| def _stub_run_result( |
| items=None, final_output="final answer text", max_turns_reached=False |
| ): |
| """Build a minimal RunResult-shaped object for serializer tests. |
| |
| Matches the shape AOB's _build_trajectory in |
| ../AssetOpsBench/src/agent/openai_agent/runner.py:121 walks. |
| """ |
| return SimpleNamespace( |
| new_items=items or [], |
| final_output=final_output, |
| max_turns_reached=max_turns_reached, |
| raw_responses=[], |
| ) |
|
|
|
|
| def test_expand_scenario_glob_accepts_shell_style_explicit_list(tmp_path: Path): |
| from scripts.aat_runner import _expand_scenario_glob |
|
|
| first = tmp_path / "multi_01_end_to_end_fault_response.json" |
| second = tmp_path / "multi_02_dga_to_workorder_pipeline.json" |
| first.write_text("{}", encoding="utf-8") |
| second.write_text("{}", encoding="utf-8") |
|
|
| scenario_files = _expand_scenario_glob(f"{first.name} {second.name}", tmp_path) |
|
|
| assert [path.name for path in scenario_files] == [first.name, second.name] |
|
|
|
|
| def test_expand_scenario_glob_accepts_single_bracket_glob(tmp_path: Path): |
| from scripts.aat_runner import _expand_scenario_glob |
|
|
| first = tmp_path / "multi_01_end_to_end_fault_response.json" |
| second = tmp_path / "multi_02_dga_to_workorder_pipeline.json" |
| third = tmp_path / "multi_03_iot_tsfm_thermal_risk.json" |
| for path in (first, second, third): |
| path.write_text("{}", encoding="utf-8") |
|
|
| scenario_files = _expand_scenario_glob("multi_0[12]_*.json", tmp_path) |
|
|
| assert [path.name for path in scenario_files] == [first.name, second.name] |
|
|
|
|
| def test_expand_scenario_glob_empty_string_returns_no_paths(tmp_path: Path): |
| from scripts.aat_runner import _expand_scenario_glob |
|
|
| assert _expand_scenario_glob("", tmp_path) == [] |
|
|
|
|
| def test_expand_scenario_glob_deduplicates_overlapping_patterns(tmp_path: Path): |
| from scripts.aat_runner import _expand_scenario_glob |
|
|
| first = tmp_path / "multi_01_end_to_end_fault_response.json" |
| second = tmp_path / "multi_02_dga_to_workorder_pipeline.json" |
| first.write_text("{}", encoding="utf-8") |
| second.write_text("{}", encoding="utf-8") |
|
|
| scenario_files = _expand_scenario_glob(f"{first.name} multi_0[12]_*.json", tmp_path) |
|
|
| assert [path.name for path in scenario_files] == [first.name, second.name] |
|
|
|
|
| def test_serialize_run_result_happy_path(): |
| from scripts.aat_runner import _serialize_run_result |
|
|
| args = _stub_args() |
| result = _stub_run_result( |
| items=[ |
| _tool_call_item("iot.list_sensors", {"asset_id": "T-1"}, "c1"), |
| _tool_output_item("[]"), |
| _msg_item("final answer text"), |
| ] |
| ) |
|
|
| out = _serialize_run_result(args, "test prompt", result, duration_seconds=12.5) |
|
|
| assert out["question"] == "test prompt" |
| assert out["answer"] == "final answer text" |
| assert out["success"] is True |
| assert out["max_turns_exhausted"] is False |
| assert out["turn_count"] >= 1 |
| assert out["tool_call_count"] == 1 |
| assert out["runner_meta"]["model_id"] == args.model_id |
| assert out["runner_meta"]["mcp_mode"] == "direct" |
| assert out["runner_meta"]["max_turns"] == 30 |
| assert out["runner_meta"]["parallel_tool_calls"] is False |
| assert out["runner_meta"]["duration_seconds"] == 12.5 |
| assert "aob_prompt_sha" in out["runner_meta"] |
|
|
| |
| |
| assert len(out["history"]) >= 1 |
| turn0 = out["history"][0] |
| assert turn0["content"] == "" |
| assert len(turn0["tool_calls"]) == 1 |
| assert turn0["tool_calls"][0]["name"] == "iot.list_sensors" |
| assert turn0["tool_calls"][0]["output"] == "[]" |
| |
| assert out["history"][-1]["content"] == "final answer text" |
|
|
|
|
| def test_serialize_run_result_warns_on_orphan_tool_output(caplog): |
| from scripts.aat_runner import _serialize_run_result |
|
|
| args = _stub_args() |
| result = _stub_run_result(items=[_tool_output_item("late output")]) |
|
|
| with caplog.at_level(logging.WARNING, logger="aat_runner"): |
| out = _serialize_run_result(args, "test prompt", result, duration_seconds=1.0) |
|
|
| assert out["success"] is True |
| assert "without a pending tool call" in caplog.text |
|
|
|
|
| def test_serialize_run_result_max_turns_exhausted(): |
| from scripts.aat_runner import _serialize_run_result |
|
|
| args = _stub_args() |
| result = _stub_run_result( |
| items=[_msg_item("still thinking")], |
| final_output=None, |
| max_turns_reached=True, |
| ) |
|
|
| out = _serialize_run_result(args, "test prompt", result, duration_seconds=99.0) |
|
|
| assert out["success"] is False |
| assert out["max_turns_exhausted"] is True |
|
|
|
|
| |
| |
| |
|
|
|
|
| def _stub_run_result_with_usage( |
| *, |
| input_tokens: int | None = 1234, |
| output_tokens: int | None = 567, |
| total_tokens: int | None = 1801, |
| requests: int | None = 5, |
| ): |
| """Stub RunResult with a context_wrapper.usage that mirrors the SDK shape.""" |
| return SimpleNamespace( |
| new_items=[_msg_item("answer")], |
| final_output="answer", |
| max_turns_reached=False, |
| raw_responses=[], |
| context_wrapper=SimpleNamespace( |
| usage=SimpleNamespace( |
| input_tokens=input_tokens, |
| output_tokens=output_tokens, |
| total_tokens=total_tokens, |
| requests=requests, |
| ) |
| ), |
| ) |
|
|
|
|
| def test_serialize_usage_captured_when_present(): |
| """Trial JSON should carry the usage block from RunContextWrapper.usage.""" |
| from scripts.aat_runner import _serialize_run_result |
|
|
| out = _serialize_run_result( |
| _stub_args(), "p", _stub_run_result_with_usage(), duration_seconds=1.0 |
| ) |
|
|
| assert out["usage"] == { |
| "input_tokens": 1234, |
| "output_tokens": 567, |
| "total_tokens": 1801, |
| "requests": 5, |
| } |
|
|
|
|
| def test_serialize_usage_missing_yields_null_block(): |
| """Older SDK / stubbed result without context_wrapper -> all-null usage block. |
| |
| summary.json's aggregator distinguishes None (no data) from 0 (zero |
| tokens), so we deliberately don't fall back to zeros here. |
| """ |
| from scripts.aat_runner import _serialize_run_result |
|
|
| out = _serialize_run_result( |
| _stub_args(), "p", _stub_run_result(), duration_seconds=1.0 |
| ) |
|
|
| assert out["usage"] == { |
| "input_tokens": None, |
| "output_tokens": None, |
| "total_tokens": None, |
| "requests": None, |
| } |
|
|
|
|
| def test_serialize_usage_negative_or_non_int_treated_as_missing(): |
| """A malformed SDK Usage with strings / negatives shouldn't poison the block.""" |
| from scripts.aat_runner import _serialize_run_result |
|
|
| out = _serialize_run_result( |
| _stub_args(), |
| "p", |
| _stub_run_result_with_usage( |
| input_tokens="not-an-int", |
| output_tokens=-1, |
| total_tokens=None, |
| requests=True, |
| ), |
| duration_seconds=1.0, |
| ) |
|
|
| assert out["usage"] == { |
| "input_tokens": None, |
| "output_tokens": None, |
| "total_tokens": None, |
| "requests": None, |
| } |
|
|
|
|
| |
| |
| |
|
|
|
|
| def test_json_default_serializes_datetime_and_date(): |
| """stdlib datetime + date land as ISO-8601 strings, matching the prior behavior.""" |
| import datetime as dt |
|
|
| from scripts.aat_runner import _json_default |
|
|
| assert _json_default(dt.datetime(2026, 4, 28, 17, 0, 0)) == "2026-04-28T17:00:00" |
| assert _json_default(dt.date(2026, 4, 28)) == "2026-04-28" |
|
|
|
|
| def test_json_default_serializes_pandas_timestamp_when_available(): |
| """pandas.Timestamp -> isoformat. Skip cleanly if pandas isn't installed.""" |
| pd = pytest.importorskip("pandas") |
| from scripts.aat_runner import _json_default |
|
|
| ts = pd.Timestamp("2026-04-28T17:00:00") |
| assert _json_default(ts) == ts.isoformat() |
|
|
|
|
| def test_json_default_serializes_numpy_datetime64_when_available(): |
| """numpy.datetime64 -> ISO string via pandas if both installed.""" |
| np = pytest.importorskip("numpy") |
| from scripts.aat_runner import _json_default |
|
|
| out = _json_default(np.datetime64("2026-04-28T17:00:00")) |
| |
| |
| assert "2026-04-28" in out |
|
|
|
|
| def test_json_default_rejects_numpy_array(): |
| """numpy.ndarray must raise TypeError, not silently stringify (Anonymous reviewer's M5).""" |
| np = pytest.importorskip("numpy") |
| from scripts.aat_runner import _json_default |
|
|
| with pytest.raises(TypeError, match="numpy"): |
| _json_default(np.array([1, 2, 3])) |
|
|
|
|
| def test_json_default_rejects_pandas_dataframe(): |
| """pandas.DataFrame must raise TypeError, not write the human-readable table.""" |
| pd = pytest.importorskip("pandas") |
| from scripts.aat_runner import _json_default |
|
|
| with pytest.raises(TypeError, match="pandas"): |
| _json_default(pd.DataFrame({"a": [1, 2, 3]})) |
|
|
|
|
| def test_json_default_rejects_arbitrary_object(): |
| """Anything outside the allow-list must raise TypeError, not stringify.""" |
| from scripts.aat_runner import _json_default |
|
|
| class Unhandled: |
| pass |
|
|
| with pytest.raises(TypeError, match="unserializable"): |
| _json_default(Unhandled()) |
|
|
|
|
| def test_write_output_uses_type_specific_encoder(tmp_path): |
| """End-to-end: _write_output rejects a payload containing a numpy array. |
| |
| The prior code would silently write '[1 2 3]'; the new code fails closed. |
| """ |
| np = pytest.importorskip("numpy") |
| from scripts.aat_runner import _write_output |
|
|
| bad_payload = {"answer": "ok", "rogue": np.array([1, 2, 3])} |
| with pytest.raises(TypeError, match="numpy"): |
| _write_output(tmp_path / "out.json", bad_payload) |
|
|
|
|
| def test_write_output_serializes_timestamp_payload(tmp_path): |
| """The Timestamp case PR #130 originally fixed must still work.""" |
| pd = pytest.importorskip("pandas") |
| from scripts.aat_runner import _write_output |
|
|
| payload = {"answer": "ok", "ts": pd.Timestamp("2026-04-28T17:00:00")} |
| out = tmp_path / "trial.json" |
| _write_output(out, payload) |
| loaded = json.loads(out.read_text()) |
| assert loaded["ts"] == "2026-04-28T17:00:00" |
|
|
|
|
| def test_cli_parses_required_args(): |
| from scripts.aat_runner import build_parser |
|
|
| parser = build_parser() |
| args = parser.parse_args( |
| [ |
| "--prompt", |
| "p", |
| "--output", |
| "/tmp/o.json", |
| "--model-id", |
| "watsonx/x", |
| "--mcp-mode", |
| "direct", |
| ] |
| ) |
| assert args.prompt == "p" |
| assert args.mcp_mode == "direct" |
| assert args.max_turns == 30 |
| assert args.parallel_tool_calls is False |
|
|
|
|
| def test_cli_parses_parallel_tool_call_modes(): |
| from scripts.aat_runner import build_parser |
|
|
| parser = build_parser() |
| common = [ |
| "--prompt", |
| "p", |
| "--output", |
| "/tmp/o.json", |
| "--model-id", |
| "watsonx/x", |
| "--mcp-mode", |
| "direct", |
| "--parallel-tool-calls", |
| ] |
|
|
| assert parser.parse_args(common + ["false"]).parallel_tool_calls is False |
| assert parser.parse_args(common + ["true"]).parallel_tool_calls is True |
| assert parser.parse_args(common + ["auto"]).parallel_tool_calls is None |
|
|
|
|
| def test_cli_missing_args_errors(): |
| from scripts.aat_runner import build_parser |
|
|
| parser = build_parser() |
| with pytest.raises(SystemExit): |
| parser.parse_args(["--prompt", "p"]) |
|
|
|
|
| def test_cli_invalid_mcp_mode_errors(): |
| from scripts.aat_runner import build_parser |
|
|
| parser = build_parser() |
| with pytest.raises(SystemExit): |
| parser.parse_args( |
| [ |
| "--prompt", |
| "p", |
| "--output", |
| "/tmp/o.json", |
| "--model-id", |
| "x", |
| "--mcp-mode", |
| "nonsense", |
| ] |
| ) |
|
|
|
|
| def test_runner_threads_litellm_env(monkeypatch): |
| from scripts.aat_runner import AaTRunner |
|
|
| captured: dict[str, object] = {} |
|
|
| class FakeLitellmModel: |
| def __init__(self, model, base_url=None, api_key=None): |
| captured["model"] = model |
| captured["base_url"] = base_url |
| captured["api_key"] = api_key |
|
|
| class FakeModelSettings: |
| def __init__(self, **kwargs): |
| self.kwargs = kwargs |
|
|
| class FakeAgent: |
| def __init__(self, **kwargs): |
| captured["agent_kwargs"] = kwargs |
|
|
| class FakeRunner: |
| @staticmethod |
| async def run(agent, prompt, max_turns): |
| captured["prompt"] = prompt |
| captured["max_turns"] = max_turns |
| return _stub_run_result() |
|
|
| monkeypatch.setenv("LITELLM_BASE_URL", "http://127.0.0.1:8000/v1") |
| monkeypatch.setenv("LITELLM_API_KEY", "dummy-vllm-not-checked") |
| monkeypatch.setitem( |
| sys.modules, |
| "agents", |
| SimpleNamespace( |
| Agent=FakeAgent, |
| ModelSettings=FakeModelSettings, |
| Runner=FakeRunner, |
| __version__="test", |
| ), |
| ) |
| monkeypatch.setitem(sys.modules, "agents.extensions", SimpleNamespace()) |
| monkeypatch.setitem(sys.modules, "agents.extensions.models", SimpleNamespace()) |
| monkeypatch.setitem( |
| sys.modules, |
| "agents.extensions.models.litellm_model", |
| SimpleNamespace(LitellmModel=FakeLitellmModel), |
| ) |
|
|
| runner = AaTRunner(model_id="openai/Llama-3.1-8B-Instruct", mcp_mode="direct") |
| asyncio.run(runner.run("hello")) |
|
|
| assert captured["model"] == "openai/Llama-3.1-8B-Instruct" |
| assert captured["base_url"] == "http://127.0.0.1:8000/v1" |
| assert captured["api_key"] == "dummy-vllm-not-checked" |
| assert captured["prompt"] == "hello" |
| assert captured["max_turns"] == 30 |
| settings = captured["agent_kwargs"]["model_settings"] |
| assert settings.kwargs["parallel_tool_calls"] is False |
|
|
|
|
| def test_watsonx_runner_omits_openai_only_model_settings(monkeypatch): |
| from scripts.aat_runner import AaTRunner |
|
|
| captured: dict[str, object] = {} |
|
|
| class FakeLitellmModel: |
| def __init__(self, model, base_url=None, api_key=None): |
| captured["model"] = model |
|
|
| class FakeModelSettings: |
| def __init__(self, **kwargs): |
| self.kwargs = kwargs |
|
|
| class FakeAgent: |
| def __init__(self, **kwargs): |
| captured["agent_kwargs"] = kwargs |
|
|
| class FakeRunner: |
| @staticmethod |
| async def run(agent, prompt, max_turns): |
| return _stub_run_result() |
|
|
| fake_litellm = SimpleNamespace(drop_params=False) |
| monkeypatch.setitem(sys.modules, "litellm", fake_litellm) |
| monkeypatch.setitem( |
| sys.modules, |
| "agents", |
| SimpleNamespace( |
| Agent=FakeAgent, |
| ModelSettings=FakeModelSettings, |
| Runner=FakeRunner, |
| __version__="test", |
| ), |
| ) |
| monkeypatch.setitem(sys.modules, "agents.extensions", SimpleNamespace()) |
| monkeypatch.setitem(sys.modules, "agents.extensions.models", SimpleNamespace()) |
| monkeypatch.setitem( |
| sys.modules, |
| "agents.extensions.models.litellm_model", |
| SimpleNamespace(LitellmModel=FakeLitellmModel), |
| ) |
|
|
| runner = AaTRunner( |
| model_id="watsonx/meta-llama/llama-3-3-70b-instruct", |
| mcp_mode="direct", |
| parallel_tool_calls=False, |
| ) |
| asyncio.run(runner.run("hello")) |
|
|
| assert captured["model"] == "watsonx/meta-llama/llama-3-3-70b-instruct" |
| assert "model_settings" not in captured["agent_kwargs"] |
| assert fake_litellm.drop_params is True |
|
|
|
|
| def test_watsonx_runner_rejects_parallel_tool_calls_true(monkeypatch): |
| from scripts.aat_runner import AaTRunner |
|
|
| monkeypatch.setitem( |
| sys.modules, |
| "agents", |
| SimpleNamespace( |
| Agent=lambda **_kwargs: None, |
| ModelSettings=lambda **_kwargs: None, |
| Runner=SimpleNamespace(run=lambda *_args, **_kwargs: None), |
| __version__="test", |
| ), |
| ) |
| monkeypatch.setitem(sys.modules, "agents.extensions", SimpleNamespace()) |
| monkeypatch.setitem(sys.modules, "agents.extensions.models", SimpleNamespace()) |
| monkeypatch.setitem( |
| sys.modules, |
| "agents.extensions.models.litellm_model", |
| SimpleNamespace(LitellmModel=lambda **_kwargs: None), |
| ) |
|
|
| runner = AaTRunner( |
| model_id="watsonx/meta-llama/llama-3-3-70b-instruct", |
| mcp_mode="direct", |
| parallel_tool_calls=True, |
| ) |
|
|
| with pytest.raises(ValueError, match="WatsonX does not support"): |
| asyncio.run(runner.run("hello")) |
|
|
|
|
| def test_batch_mode_requires_output_dir(): |
| from scripts.aat_runner import _main |
|
|
| args = argparse.Namespace( |
| scenarios_glob="data/scenarios/*.json", |
| output_dir=None, |
| prompt=None, |
| output=None, |
| model_id="x", |
| mcp_mode="optimized", |
| max_turns=30, |
| parallel_tool_calls=False, |
| verbose=False, |
| trials=1, |
| run_basename="batch", |
| ) |
| assert asyncio.run(_main(args)) == 2 |
|
|
|
|
| def test_batch_mode_rejects_non_optimized_mcp_mode(tmp_path): |
| from scripts.aat_runner import _main |
|
|
| args = argparse.Namespace( |
| scenarios_glob="nonexistent_xyzzy_*.json", |
| output_dir=str(tmp_path), |
| prompt=None, |
| output=None, |
| model_id="x", |
| mcp_mode="direct", |
| max_turns=30, |
| parallel_tool_calls=False, |
| verbose=False, |
| trials=1, |
| run_basename="batch", |
| ) |
| assert asyncio.run(_main(args)) == 2 |
|
|
|
|
| def test_batch_mode_rejects_zero_trials(tmp_path): |
| """--trials must be >= 1; zero/negative must return rc=2 without creating output dir.""" |
| from scripts.aat_runner import _main_multi |
| from pathlib import Path |
|
|
| repo_root = Path(__file__).resolve().parent.parent |
| new_subdir = tmp_path / "should_not_be_created" |
| for bad_trials in (0, -1): |
| args = argparse.Namespace( |
| scenarios_glob="nonexistent_xyzzy_*.json", |
| output_dir=str(new_subdir), |
| model_id="x", |
| mcp_mode="optimized", |
| max_turns=30, |
| parallel_tool_calls=False, |
| trials=bad_trials, |
| run_basename="batch", |
| ) |
| rc = asyncio.run(_main_multi(args, repo_root)) |
| assert rc == 2, f"expected rc=2 for trials={bad_trials}, got {rc}" |
| assert ( |
| not new_subdir.exists() |
| ), f"output dir must not be created for trials={bad_trials}" |
|
|
|
|
| def test_batch_output_dir_normalization(): |
| """Regression: relative --output-dir must not raise ValueError in relative_to(). |
| |
| Directly exercises the path arithmetic that crashed before the fix, without |
| calling _main_multi (which would mkdir under the repo and never reach |
| relative_to due to the empty-glob early-return). |
| """ |
| from pathlib import Path |
|
|
| repo_root = Path(__file__).resolve().parent.parent |
| |
| raw = Path("benchmarks/cell_C_mcp_optimized/raw/test-run") |
| output_dir = repo_root / raw |
| out_path = output_dir / "batch_scenario_run01.json" |
| |
| rel = out_path.relative_to(repo_root) |
| assert rel == Path( |
| "benchmarks/cell_C_mcp_optimized/raw/test-run/batch_scenario_run01.json" |
| ) |
|
|
|
|
| def test_batch_mode_rejects_non_optimized_before_mkdir(tmp_path): |
| """M1 regression: mcp_mode guard must fire before output_dir is created.""" |
| from scripts.aat_runner import _main_multi |
| from pathlib import Path |
|
|
| repo_root = Path(__file__).resolve().parent.parent |
| new_subdir = tmp_path / "should_not_be_created" |
| args = argparse.Namespace( |
| scenarios_glob="nonexistent_xyzzy_*.json", |
| output_dir=str(new_subdir), |
| model_id="x", |
| mcp_mode="direct", |
| max_turns=30, |
| parallel_tool_calls=False, |
| trials=1, |
| run_basename="batch", |
| ) |
| rc = asyncio.run(_main_multi(args, repo_root)) |
| assert rc == 2 |
| assert not new_subdir.exists() |
|
|
|
|
| def test_batch_latency_path_outside_repo_root(tmp_path): |
| """M3 regression: relative_to() fallback must not raise for absolute out-of-repo path.""" |
| from pathlib import Path |
|
|
| repo_root = Path(__file__).resolve().parent.parent |
| out_path = tmp_path / "trial_01.json" |
| |
| try: |
| result = out_path.relative_to(repo_root).as_posix() |
| except ValueError: |
| result = out_path.as_posix() |
| |
| assert result == out_path.as_posix() |
|
|