UnitCommitment_Trajectory / benchmark /scripts /download_matpower_instances.py
EridanusQ
init
43c68a3
from __future__ import annotations
import argparse
import datetime as dt
import os
import re
import sys
import time
import urllib.error
import urllib.request
from concurrent.futures import ThreadPoolExecutor, as_completed
from dataclasses import dataclass
from pathlib import Path
BASE_URL_DEFAULT = "https://axavier.org/UnitCommitment.jl/0.4/instances"
@dataclass(frozen=True)
class Task:
case: str
date: dt.date
url: str
out_path: Path
def _iter_dates(start: dt.date, end: dt.date):
if end < start:
raise ValueError("end_date must be >= start_date")
cur = start
one = dt.timedelta(days=1)
while cur <= end:
yield cur
cur += one
def _default_instances_md_path() -> Path:
here = Path(__file__).resolve()
project_root = here.parents[2]
return project_root / "docs" / "src" / "guides" / "instances.md"
def _parse_matpower_cases_from_instances_md(instances_md_path: Path) -> list[str]:
text = instances_md_path.read_text(encoding="utf-8")
pattern = re.compile(r"matpower/(case[^/\s`]+)/\d{4}-\d{2}-\d{2}")
cases = sorted(set(pattern.findall(text)))
if not cases:
raise RuntimeError(f"No MATPOWER cases found in {instances_md_path}")
return cases
def _download_one(task: Task, timeout_s: float, retries: int, force: bool) -> tuple[str, str | None]:
task.out_path.parent.mkdir(parents=True, exist_ok=True)
if task.out_path.exists() and not force and task.out_path.stat().st_size > 0:
return ("skipped", None)
tmp_path = task.out_path.with_suffix(task.out_path.suffix + ".part")
req = urllib.request.Request(task.url, headers={"User-Agent": "UnitCommitment downloader"})
last_err = None
for attempt in range(retries + 1):
try:
with urllib.request.urlopen(req, timeout=timeout_s) as resp:
with open(tmp_path, "wb") as f:
while True:
chunk = resp.read(1024 * 256)
if not chunk:
break
f.write(chunk)
os.replace(tmp_path, task.out_path)
return ("downloaded", None)
except (urllib.error.URLError, urllib.error.HTTPError, TimeoutError, OSError) as e:
last_err = e
try:
if tmp_path.exists():
tmp_path.unlink()
except OSError:
pass
if attempt < retries:
time.sleep(min(10.0, 0.5 * (2**attempt)))
continue
return ("failed", f"{type(last_err).__name__}: {last_err}")
return ("failed", f"{type(last_err).__name__}: {last_err}")
def _build_tasks(
base_url: str,
out_dir: Path,
cases: list[str],
start_date: dt.date,
end_date: dt.date,
) -> list[Task]:
tasks: list[Task] = []
for case in cases:
for d in _iter_dates(start_date, end_date):
date_str = d.isoformat()
url = f"{base_url}/matpower/{case}/{date_str}.json.gz"
out_path = out_dir / case / f"{date_str}.json.gz"
tasks.append(Task(case=case, date=d, url=url, out_path=out_path))
return tasks
def _parse_date(s: str) -> dt.date:
return dt.date.fromisoformat(s)
def main(argv: list[str]) -> int:
parser = argparse.ArgumentParser()
parser.add_argument("--instances-md", type=Path, default=_default_instances_md_path())
parser.add_argument("--base-url", type=str, default=BASE_URL_DEFAULT)
parser.add_argument(
"--out-dir",
type=Path,
default=(Path(__file__).resolve().parents[2] / "instances" / "matpower"),
)
parser.add_argument("--start-date", type=_parse_date, default=dt.date(2017, 1, 1))
parser.add_argument("--end-date", type=_parse_date, default=dt.date(2017, 12, 31))
parser.add_argument("--workers", type=int, default=min(32, (os.cpu_count() or 4) * 4))
parser.add_argument("--timeout", type=float, default=60.0)
parser.add_argument("--retries", type=int, default=3)
parser.add_argument("--force", action="store_true")
parser.add_argument("--dry-run", action="store_true")
parser.add_argument("--list-cases", action="store_true")
args = parser.parse_args(argv)
instances_md_path: Path = args.instances_md
if not instances_md_path.exists():
raise FileNotFoundError(str(instances_md_path))
cases = _parse_matpower_cases_from_instances_md(instances_md_path)
if args.list_cases:
for c in cases:
print(c)
return 0
out_dir: Path = args.out_dir
tasks = _build_tasks(
base_url=args.base_url.rstrip("/"),
out_dir=out_dir,
cases=cases,
start_date=args.start_date,
end_date=args.end_date,
)
print(f"cases={len(cases)} files={len(tasks)} out_dir={out_dir}")
if args.dry_run:
for t in tasks[:10]:
print(f"{t.url} -> {t.out_path}")
if len(tasks) > 10:
print("...")
return 0
downloaded = 0
skipped = 0
failed = 0
total = len(tasks)
t0 = time.time()
last_print = 0.0
with ThreadPoolExecutor(max_workers=max(1, args.workers)) as ex:
fut_to_task = {
ex.submit(_download_one, t, args.timeout, args.retries, args.force): t for t in tasks
}
for i, fut in enumerate(as_completed(fut_to_task), start=1):
status, err = fut.result()
if status == "downloaded":
downloaded += 1
elif status == "skipped":
skipped += 1
else:
failed += 1
task = fut_to_task[fut]
sys.stderr.write(f"FAILED {task.url} -> {task.out_path} ({err})\n")
now = time.time()
if now - last_print >= 1.0 or i == total:
elapsed = max(0.001, now - t0)
rate = i / elapsed
print(
f"{i}/{total} downloaded={downloaded} skipped={skipped} failed={failed} rate={rate:.1f}/s"
)
last_print = now
if failed:
return 2
return 0
if __name__ == "__main__":
raise SystemExit(main(sys.argv[1:]))