smartgridbench-review-artifact / code /tests /test_gcp_resume_state.py
garn-garn's picture
Add anonymized executable code package
881f9f2 verified
from __future__ import annotations
import json
from pathlib import Path
from scripts import gcp_resume_state
def _scenario(path: Path, scenario_id: str = "SGT-001") -> Path:
path.write_text(
json.dumps({"id": scenario_id, "text": "test prompt"}) + "\n",
encoding="utf-8",
)
return path
def _trial(path: Path, *, success: bool | None = True, answer: str = "done") -> Path:
payload = {"answer": answer, "history": []}
if success is not None:
payload["success"] = success
path.parent.mkdir(parents=True, exist_ok=True)
path.write_text(json.dumps(payload) + "\n", encoding="utf-8")
return path
def _latency(
path: Path, scenario_file: Path, trial_index: int, output_path: Path
) -> None:
path.parent.mkdir(parents=True, exist_ok=True)
with path.open("a", encoding="utf-8") as fh:
fh.write(
json.dumps(
{
"scenario_file": scenario_file.as_posix(),
"trial_index": trial_index,
"latency_seconds": 1.25,
"output_path": output_path.as_posix(),
}
)
+ "\n"
)
def _classify(
tmp_path: Path,
scenario_file: Path,
output_path: Path,
*,
require_latency: bool = True,
) -> dict:
return gcp_resume_state.classify_trial(
run_dir=tmp_path,
scenario_file=scenario_file,
trial_index=1,
output_path=output_path,
latency_file=tmp_path / "latencies.jsonl",
require_latency=require_latency,
)
def test_completed_success_is_terminal_with_latency(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = _trial(tmp_path / "2026-05-03_Y_multi_01_run01.json", success=True)
_latency(tmp_path / "latencies.jsonl", scenario_file, 1, output)
result = _classify(tmp_path, scenario_file, output)
assert result["state"] == "complete_success"
assert result["complete"] is True
assert result["success"] is True
def test_completed_failure_is_terminal_and_skippable(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = _trial(tmp_path / "2026-05-03_Y_multi_01_run01.json", success=False)
_latency(tmp_path / "latencies.jsonl", scenario_file, 1, output)
result = _classify(tmp_path, scenario_file, output)
assert result["state"] == "complete_failure"
assert result["complete"] is True
assert result["success"] is False
def test_missing_latency_keeps_valid_json_incomplete_by_default(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = _trial(tmp_path / "2026-05-03_Y_multi_01_run01.json", success=True)
result = _classify(tmp_path, scenario_file, output)
assert result["state"] == "incomplete"
assert "missing_latency" in result["reason"]
def test_invalid_json_is_incomplete(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = tmp_path / "2026-05-03_Y_multi_01_run01.json"
output.write_text("{not-json", encoding="utf-8")
result = _classify(tmp_path, scenario_file, output, require_latency=False)
assert result["state"] == "incomplete"
assert "invalid_json" in result["reason"]
def test_legacy_stdout_next_to_terminal_json_is_not_dangling(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = _trial(tmp_path / "legacy_multi_01_run01.json", success=True)
output.with_suffix(output.suffix + ".stdout").write_text(
"legacy log\n", encoding="utf-8"
)
_latency(tmp_path / "latencies.jsonl", scenario_file, 1, output)
newer_name = tmp_path / "2026-05-04_Y_multi_01_run01.json"
result = _classify(tmp_path, scenario_file, newer_name)
assert result["state"] == "complete_success"
assert result["output_path"] == output.as_posix()
def test_conflicting_duplicate_latency_rows_are_incomplete(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = _trial(tmp_path / "2026-05-03_Y_multi_01_run01.json", success=True)
rows = [
{
"scenario_file": scenario_file.as_posix(),
"trial_index": 1,
"latency_seconds": 1.0,
"output_path": output.as_posix(),
},
{
"scenario_file": scenario_file.as_posix(),
"trial_index": 1,
"latency_seconds": 2.0,
"output_path": output.as_posix(),
},
]
(tmp_path / "latencies.jsonl").write_text(
"".join(json.dumps(row) + "\n" for row in rows),
encoding="utf-8",
)
result = _classify(tmp_path, scenario_file, output)
assert result["state"] == "incomplete"
assert "duplicate_conflicting_latency" in result["reason"]
def test_finalize_trial_writes_atomic_output_latency_and_manifest(
tmp_path: Path,
) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
temp_output = _trial(tmp_path / "run" / "trial.json.tmp", success=None)
output = tmp_path / "run" / "trial.json"
latency_file = tmp_path / "run" / "latencies.jsonl"
manifest_file = tmp_path / "run" / "resume_manifest.jsonl"
result = gcp_resume_state.finalize_trial(
scenario_file=scenario_file,
trial_index=1,
temp_output=temp_output,
output_path=output,
latency_file=latency_file,
manifest_file=manifest_file,
run_name="resume-smoke",
batch_id="batch-1",
start_epoch=10.0,
end_epoch=12.5,
return_code=0,
)
assert result["state"] == "complete_success"
assert output.exists()
assert not temp_output.exists()
payload = json.loads(output.read_text(encoding="utf-8"))
assert payload["scenario"]["id"] == "SGT-001"
assert payload["success"] is True
latency_rows = [json.loads(line) for line in latency_file.read_text().splitlines()]
assert latency_rows == [
{
"latency_seconds": 2.5,
"output_path": output.as_posix(),
"scenario_file": scenario_file.as_posix(),
"trial_index": 1,
}
]
manifest_rows = [
json.loads(line)
for line in manifest_file.read_text(encoding="utf-8").splitlines()
]
assert manifest_rows[-1]["state"] == "complete_success"
assert manifest_rows[-1]["batch_id"] == "batch-1"
assert "runtime_versions" in manifest_rows[-1]
def test_finalize_trial_marks_divergent_rerun(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
output = tmp_path / "run" / "trial.json"
latency_file = tmp_path / "run" / "latencies.jsonl"
manifest_file = tmp_path / "run" / "resume_manifest.jsonl"
first_temp = _trial(tmp_path / "run" / "trial-first.json.tmp", answer="first")
second_temp = _trial(tmp_path / "run" / "trial-second.json.tmp", answer="second")
gcp_resume_state.finalize_trial(
scenario_file=scenario_file,
trial_index=1,
temp_output=first_temp,
output_path=output,
latency_file=latency_file,
manifest_file=manifest_file,
run_name="resume-smoke",
batch_id="batch-1",
start_epoch=10.0,
end_epoch=12.5,
return_code=0,
)
gcp_resume_state.finalize_trial(
scenario_file=scenario_file,
trial_index=1,
temp_output=second_temp,
output_path=output,
latency_file=latency_file,
manifest_file=manifest_file,
run_name="resume-smoke",
batch_id="batch-1",
start_epoch=20.0,
end_epoch=23.0,
return_code=0,
)
manifest_rows = [
json.loads(line)
for line in manifest_file.read_text(encoding="utf-8").splitlines()
]
rerun = manifest_rows[-1]
assert rerun["divergent"] is True
assert rerun["original_output_sha256"] != rerun["final_output_sha256"]
def test_validate_run_artifacts_flags_missing_outputs(tmp_path: Path) -> None:
scenario_file = _scenario(tmp_path / "multi_01.json")
result = gcp_resume_state.validate_run_artifacts(
run_dir=tmp_path / "run",
scenario_glob=scenario_file.as_posix(),
trials=1,
run_name="batch_Y",
require_latency=True,
)
assert result["valid"] is False
assert result["expected"] == 1
assert result["complete"] == 0
assert result["reason"] == "incomplete_trajectory_artifacts"
assert result["missing"][0]["reason"] == "missing_json"
def test_validate_run_artifacts_counts_terminal_failures_as_complete(
tmp_path: Path,
) -> None:
scenarios_dir = tmp_path / "scenarios"
scenarios_dir.mkdir()
scenario_file = _scenario(scenarios_dir / "multi_01.json")
run_dir = tmp_path / "run"
run_name = "batch_Y"
success_output = _trial(
run_dir / f"{run_name}_multi_01_run01.json",
success=True,
)
failure_output = _trial(
run_dir / f"{run_name}_multi_01_run02.json",
success=False,
)
_latency(run_dir / "latencies.jsonl", scenario_file, 1, success_output)
_latency(run_dir / "latencies.jsonl", scenario_file, 2, failure_output)
result = gcp_resume_state.validate_run_artifacts(
run_dir=run_dir,
scenario_glob=scenario_file.as_posix(),
trials=2,
run_name=run_name,
require_latency=True,
)
assert result["valid"] is True
assert result["expected"] == 2
assert result["complete"] == 2
assert result["success"] == 1
assert result["failure"] == 1
assert result["missing"] == []
def test_validate_run_artifacts_accepts_space_separated_scenario_list(
tmp_path: Path,
) -> None:
scenarios_dir = tmp_path / "scenarios"
scenarios_dir.mkdir()
first_scenario = _scenario(scenarios_dir / "multi_01.json", "SGT-001")
second_scenario = _scenario(scenarios_dir / "wo_04.json", "SGT-018")
run_dir = tmp_path / "run"
run_name = "batch_Y"
first_output = _trial(
run_dir / f"{run_name}_multi_01_run01.json",
success=True,
)
second_output = _trial(
run_dir / f"{run_name}_wo_04_run01.json",
success=True,
)
_latency(run_dir / "latencies.jsonl", first_scenario, 1, first_output)
_latency(run_dir / "latencies.jsonl", second_scenario, 1, second_output)
result = gcp_resume_state.validate_run_artifacts(
run_dir=run_dir,
scenario_glob=f"{first_scenario.as_posix()} {second_scenario.as_posix()}",
trials=1,
run_name=run_name,
require_latency=True,
)
assert result["valid"] is True
assert result["expected"] == 2
assert result["complete"] == 2
assert result["missing"] == []