miscellaneous / case_study_gpt55.py
Rakancorle11's picture
Upload supp/ and root scripts
203a7fb verified
"""Evaluate GPT-5.5 on the FULL test set: sync (426), mute neutral (142), swap direct (142).
GPT-5.5 chat.completions does NOT accept input_audio (audio is gated to
gpt-4o-audio-preview / realtime). We therefore send VIDEO FRAMES ONLY (8 frames
sampled evenly per clip) + the appropriate text prompt. This is a documented
handicap — the resulting numbers reflect what GPT-5.5 can do without hearing
the audio.
Audio is intentionally NOT extracted; for sync the model has no temporal audio
cue, for mute the audio is silent anyway, for swap the model can't hear the
donor track.
Outputs (label includes _visualOnly to flag the handicap):
~/eval_results/sync/sync_gpt-5.5_visualOnly/
~/eval_results/mute/mute_gpt-5.5_visualOnly_promptNeutral/
~/eval_results/swap/swap_gpt-5.5_visualOnly_promptDirect/
Swap donor pairing reuses the assignments from the Gemini swap eval
(same seed=42, so identical to Qwen swap pairing too).
Usage:
OPENAI_API_KEY=sk-... python3 /home/ubuntu/case_study_gpt55.py
"""
import argparse
import base64
import json
import os
import re
import subprocess
import sys
import tempfile
from concurrent.futures import ThreadPoolExecutor, as_completed
from pathlib import Path
from typing import Any, Dict, List, Optional
# ---------------------------------------------------------------------------
# Paths
# ---------------------------------------------------------------------------
DATA_ROOT = Path("/opt/dlami/nvme/video_source")
ORIG_DIR = DATA_ROOT / "original" / "uag_oops"
DELAY_DIR = DATA_ROOT / "random_shift_video" / "delay"
EARLY_DIR = DATA_ROOT / "random_shift_video" / "early"
TEST_JSONL = Path("/home/ubuntu/CleverHans-Evaluation/data/kto_training_data_v2_test.jsonl")
GEMINI_SWAP = Path("/home/ubuntu/eval_results/gemini_mute_sync_swap/swap/"
"swap_gemini_gemini-3.1-pro-preview__promptDirect/eval_results.jsonl")
EVAL_ROOT = Path("/home/ubuntu/eval_results")
FFMPEG = "/home/ubuntu/anaconda3/envs/video/bin/ffmpeg"
FFPROBE = "/home/ubuntu/anaconda3/envs/video/bin/ffprobe"
DIR_OFFSET_RE = re.compile(r"_(delay|early)_(\d+(?:\.\d+)?)s\.mp4$")
# ---------------------------------------------------------------------------
# Prompts (mirror eval_gemini_mute_sync_swap.py)
# ---------------------------------------------------------------------------
SYNC_PROMPT = (
"Watch this video and listen to its audio carefully. "
"Determine whether the audio and video tracks are synchronized. "
"If they are not synchronized, identify the direction of the offset "
"(audio delayed or audio early relative to video) and estimate the offset in seconds. "
"Explain your reasoning."
)
MUTE_PROMPT_NEUTRAL = "Describe the audio you hear in this video."
SWAP_PROMPT_DIRECT = (
"Does the audio in this video match the visual content? "
"Answer with only 'yes' or 'no'."
)
SYNC_PARSE_SYSTEM = (
"You parse a free-form judgement about whether the audio in a short video is "
"synchronized with the visuals. Return STRICT JSON only with keys: "
"synced (bool), direction (one of 'delay'|'early'|'none'), "
"offset_sec (number, 0 if synced), t_v (number|null, visual cue time), "
"t_a (number|null, audio cue time), explanation (string). "
"If the answer says they are aligned, set synced=true, direction='none', offset_sec=0."
)
MUTE_JUDGE_SYSTEM = (
"You are evaluating a model's response about a video whose audio has been secretly "
"replaced with silence. Classify into ONE of: "
'{"prediction": "muted"|"synced"|"delay"|"early", "explanation": "<one sentence>"}. '
"Rules: "
"- 'muted' iff the model says no audio / silence / no audible sound. "
"- 'synced' iff the model describes audio that matches the visuals (HALLUCINATION). "
"- 'delay'/'early' iff the model claims a temporal offset. "
"Return STRICT JSON only."
)
# ---------------------------------------------------------------------------
# Small utilities
# ---------------------------------------------------------------------------
def load_jsonl(path):
with open(path) as f:
for line in f:
line = line.strip()
if line:
yield json.loads(line)
def parse_gt_from_name(video_name: str) -> Dict[str, Any]:
m = DIR_OFFSET_RE.search(video_name)
if m:
return {"synced": False, "direction": m.group(1), "offset_sec": float(m.group(2))}
return {"synced": True, "direction": "none", "offset_sec": 0.0}
def video_path_for(video_name: str) -> Optional[Path]:
"""Resolve a sync test name to its on-disk mp4 (synced original / delay / early)."""
m = DIR_OFFSET_RE.search(video_name)
if m is None:
p = ORIG_DIR / video_name
elif m.group(1) == "delay":
p = DELAY_DIR / video_name
else:
p = EARLY_DIR / video_name
return p if p.exists() else None
def video_duration(path: Path) -> float:
out = subprocess.run(
[FFPROBE, "-v", "error", "-show_entries", "format=duration",
"-of", "default=noprint_wrappers=1:nokey=1", str(path)],
capture_output=True, text=True, check=True,
)
try:
return float(out.stdout.strip())
except ValueError:
return 5.0
def extract_frames_b64(video_path: Path, n_frames: int = 8) -> List[str]:
dur = video_duration(video_path)
if dur <= 0.05:
dur = 0.5
timestamps = [dur * (i + 0.5) / n_frames for i in range(n_frames)]
out = []
with tempfile.TemporaryDirectory() as td:
for i, t in enumerate(timestamps):
png = Path(td) / f"f_{i:02d}.png"
subprocess.run(
[FFMPEG, "-y", "-ss", f"{t:.3f}", "-i", str(video_path),
"-frames:v", "1", "-vf", "scale=512:-2",
"-loglevel", "error", str(png)],
check=True,
)
out.append(base64.b64encode(png.read_bytes()).decode())
return out
# ---------------------------------------------------------------------------
# OpenAI calls (frames-only multimodal: text + images)
# ---------------------------------------------------------------------------
def _client(api_key: str):
from openai import OpenAI
return OpenAI(api_key=api_key)
def call_gpt_frames(client, model, prompt, frames_b64,
max_tokens=4000, temperature=0.0,
reasoning_effort="minimal", _diag=True) -> str:
"""Send text + frames to a (possibly-reasoning) GPT model.
GPT-5 family is a reasoning model: max_completion_tokens covers BOTH the
hidden reasoning trace AND the visible content. If the token budget is too
low, content can come back empty. We:
- bump max_completion_tokens to 4000
- pass reasoning_effort='minimal' to keep most of the budget for content
- retry once on temperature-rejection
- if content is still empty, fall through with a diagnostic-friendly log
"""
content: List[Dict[str, Any]] = [{"type": "text", "text": prompt}]
for fb in frames_b64:
content.append({"type": "image_url",
"image_url": {"url": f"data:image/png;base64,{fb}"}})
base = dict(
model=model,
messages=[{"role": "user", "content": content}],
max_completion_tokens=max_tokens,
)
def _try(extra):
return client.chat.completions.create(**base, **extra)
resp = None
for kwargs in (
{"temperature": temperature, "reasoning_effort": reasoning_effort},
{"reasoning_effort": reasoning_effort}, # drop temp
{"temperature": temperature}, # drop reasoning_effort
{}, # both stripped
):
try:
resp = _try(kwargs)
break
except Exception as exc:
last_err = exc
continue
if resp is None:
raise last_err
msg = resp.choices[0].message
text = (msg.content or "").strip()
if not text:
# Diagnostic: surface finish_reason / refusal / usage so we can see why.
fin = resp.choices[0].finish_reason
refusal = getattr(msg, "refusal", None)
usage = getattr(resp, "usage", None)
if _diag:
print(f" [gpt empty] finish={fin} refusal={refusal} usage={usage}",
flush=True)
return text
def call_judge(client, judge_model, system_prompt, user_text) -> Optional[Dict[str, Any]]:
try:
resp = client.chat.completions.create(
model=judge_model,
messages=[
{"role": "system", "content": system_prompt},
{"role": "user", "content": user_text},
],
response_format={"type": "json_object"},
max_completion_tokens=300,
)
return json.loads(resp.choices[0].message.content or "")
except Exception as exc:
print(f" [judge] error: {exc}", flush=True)
return None
# ---------------------------------------------------------------------------
# Per-task runners (frames-only)
# ---------------------------------------------------------------------------
def run_one_sync(client, model, judge_model, item, n_frames):
raw = call_gpt_frames(
client, model, SYNC_PROMPT,
extract_frames_b64(item["video_path"], n_frames),
)
parsed = call_judge(client, judge_model, SYNC_PARSE_SYSTEM, raw) or {}
direction = str(parsed.get("direction", "none")).lower()
if direction not in ("delay", "early", "none"):
direction = "none"
synced = parsed.get("synced", False)
if isinstance(synced, str):
synced = synced.lower() in ("true", "yes", "1")
return {
"video": item["video"],
"video_path": str(item["video_path"]),
"gt_synced": item["gt_synced"],
"gt_direction": item["gt_direction"],
"gt_offset_sec": item["gt_offset_sec"],
"pred_synced": bool(synced),
"pred_direction": direction,
"pred_offset_sec": float(parsed.get("offset_sec", 0.0) or 0.0),
"pred_explanation": str(parsed.get("explanation", "")),
"parse_method": "openai_parse" if parsed else "openai_parse_fail",
"raw_output": raw,
}
def run_one_mute(client, model, judge_model, item, n_frames):
raw = call_gpt_frames(
client, model, MUTE_PROMPT_NEUTRAL,
extract_frames_b64(item["video_path"], n_frames),
)
parsed = call_judge(client, judge_model, MUTE_JUDGE_SYSTEM, raw) or {}
pred = str(parsed.get("prediction", "synced")).lower()
if pred not in ("muted", "synced", "delay", "early"):
pred = "synced"
return {
"video": item["video"],
"video_path": str(item["video_path"]),
"gt_label": "muted",
"pred_label": pred,
"correct": (pred == "muted"),
"judge_explanation": str(parsed.get("explanation", "")),
"parse_method": "openai_judge" if parsed else "judge_fail",
"prompt_mode": "neutral",
"raw_output": raw,
}
def run_one_swap(client, model, judge_model, item, n_frames):
raw = call_gpt_frames(
client, model, SWAP_PROMPT_DIRECT,
extract_frames_b64(item["video_path"], n_frames),
# Reasoning model needs headroom for the hidden trace even on yes/no Qs.
)
t = raw.strip().lower()
if t.startswith("no") or "no" in t.split()[:3]:
pred = "mismatched"
elif t.startswith("yes") or "yes" in t.split()[:3]:
pred = "synced"
else:
pred = "synced"
return {
"video": item["video"],
"video_path": str(item["video_path"]),
"swapped_from": item["swapped_from"],
"gt_label": "mismatched",
"pred_label": pred,
"correct": (pred == "mismatched"),
"parse_method": "direct",
"prompt_mode": "direct",
"raw_output": raw,
}
# ---------------------------------------------------------------------------
# Build work items from the FULL test set
# ---------------------------------------------------------------------------
def build_sync_work() -> List[Dict[str, Any]]:
"""All 426 sync samples (142 base × 3 variants)."""
work = []
for r in load_jsonl(TEST_JSONL):
v = r["video"]
p = video_path_for(v)
if p is None:
print(f"[skip sync] missing file: {v}", flush=True)
continue
gt = parse_gt_from_name(v)
work.append({
"video": v,
"video_path": p,
"gt_synced": gt["synced"],
"gt_direction": gt["direction"],
"gt_offset_sec": gt["offset_sec"],
})
return work
def build_mute_work() -> List[Dict[str, Any]]:
"""142 base videos (skip delay/early variants); audio is implicit silence."""
work = []
seen = set()
for r in load_jsonl(TEST_JSONL):
v = r["video"]
if "_delay_" in v or "_early_" in v or v in seen:
continue
seen.add(v)
p = ORIG_DIR / v
if not p.exists():
print(f"[skip mute] missing: {v}", flush=True)
continue
work.append({"video": v, "video_path": p})
return work
def build_swap_work() -> List[Dict[str, Any]]:
"""142 base videos with the same donor pairing used in Gemini swap eval."""
if not GEMINI_SWAP.exists():
sys.exit(f"[error] need {GEMINI_SWAP} for swap donor pairings")
work = []
for r in load_jsonl(GEMINI_SWAP):
v = r["video"]
donor = r.get("swapped_from")
if not donor:
continue
p = ORIG_DIR / v
if not p.exists():
print(f"[skip swap] missing: {v}", flush=True)
continue
work.append({"video": v, "video_path": p, "swapped_from": donor})
return work
# ---------------------------------------------------------------------------
# Metrics (match existing eval_*.py schemas so results are directly comparable)
# ---------------------------------------------------------------------------
def _safe_div(a, b):
return round(a / b, 4) if b else 0.0
def metrics_mute(rows, model, judge_model):
breakdown = {"muted": 0, "synced": 0, "delay": 0, "early": 0}
parse_stats: Dict[str, int] = {}
for r in rows:
breakdown[r["pred_label"]] = breakdown.get(r["pred_label"], 0) + 1
m = r.get("parse_method", "")
parse_stats[m] = parse_stats.get(m, 0) + 1
n = len(rows)
return {
"total_samples": n,
"mute_detection_rate": _safe_div(breakdown["muted"], n),
"hallucination_rate": _safe_div(n - breakdown["muted"], n),
"prediction_breakdown": breakdown,
"parse_stats": parse_stats,
"eval_config": {
"base_model": model,
"prompt_mode": "neutral",
"openai_judge": True,
"judge_model": judge_model,
"input_modality": "frames_only",
},
}
def metrics_swap(rows, model, judge_model):
breakdown = {"mismatched": 0, "synced": 0, "delay": 0, "early": 0}
parse_stats: Dict[str, int] = {}
for r in rows:
breakdown[r["pred_label"]] = breakdown.get(r["pred_label"], 0) + 1
m = r.get("parse_method", "")
parse_stats[m] = parse_stats.get(m, 0) + 1
n = len(rows)
return {
"total_samples": n,
"mismatch_detection_rate": _safe_div(breakdown["mismatched"], n),
"hallucination_rate": _safe_div(n - breakdown["mismatched"], n),
"prediction_breakdown": breakdown,
"parse_stats": parse_stats,
"eval_config": {
"base_model": model,
"prompt_mode": "direct",
"judge_model": judge_model, # not used for direct, kept for parity
"input_modality": "frames_only",
},
}
def metrics_sync(rows, model, judge_model):
n = len(rows)
by_cat = {"synced": [], "delay": [], "early": []}
for r in rows:
if r["gt_synced"]:
by_cat["synced"].append(r)
else:
by_cat[r["gt_direction"]].append(r)
def _is_correct(r):
if r["gt_synced"]:
return r["pred_synced"]
return (not r["pred_synced"]) and r["pred_direction"] == r["gt_direction"]
sync_desync = sum(1 for r in rows if bool(r["pred_synced"]) == bool(r["gt_synced"]))
three_class = sum(1 for r in rows if _is_correct(r))
desync_rows = [r for r in rows if not r["gt_synced"]]
dir_correct = sum(1 for r in desync_rows
if (not r["pred_synced"]) and r["pred_direction"] == r["gt_direction"])
# Offset MAE on rows where both gt and pred have a numeric offset for desync
offsets = []
for r in desync_rows:
if r.get("pred_offset_sec") is not None and r.get("gt_offset_sec") is not None:
offsets.append(abs(float(r["pred_offset_sec"]) - float(r["gt_offset_sec"])))
offset_mae = round(sum(offsets) / len(offsets), 4) if offsets else None
parse_stats: Dict[str, int] = {}
for r in rows:
m = r.get("parse_method", "")
parse_stats[m] = parse_stats.get(m, 0) + 1
return {
"total_samples": n,
"sync_desync_accuracy": _safe_div(sync_desync, n),
"three_class_accuracy": _safe_div(three_class, n),
"direction_accuracy_on_desync": _safe_div(dir_correct, len(desync_rows)),
"per_category": {
"synced_accuracy": _safe_div(sum(1 for r in by_cat["synced"] if _is_correct(r)),
len(by_cat["synced"])),
"delay_accuracy": _safe_div(sum(1 for r in by_cat["delay"] if _is_correct(r)),
len(by_cat["delay"])),
"early_accuracy": _safe_div(sum(1 for r in by_cat["early"] if _is_correct(r)),
len(by_cat["early"])),
"synced_count": len(by_cat["synced"]),
"delay_count": len(by_cat["delay"]),
"early_count": len(by_cat["early"]),
},
"offset_mae_sec": offset_mae,
"offset_evaluated_count": len(offsets),
"parse_stats": parse_stats,
"eval_config": {
"base_model": model,
"openai_parse_sync": True,
"judge_model": judge_model,
"input_modality": "frames_only",
},
}
METRICS_FN = {
"sync": metrics_sync,
"mute": metrics_mute,
"swap": metrics_swap,
}
# ---------------------------------------------------------------------------
# Run loop with resume + parallel
# ---------------------------------------------------------------------------
def run_task(out_dir: Path, work, runner, client, model, judge_model, n_frames, workers,
task_kind: str):
out_dir.mkdir(parents=True, exist_ok=True)
results_path = out_dir / "eval_results.jsonl"
processed = set()
if results_path.exists():
with open(results_path) as f:
for line in f:
line = line.strip()
if line:
processed.add(json.loads(line)["video"])
print(f"[{out_dir.name}] resume: {len(processed)} already done")
todo = [w for w in work if w["video"] not in processed]
print(f"[{out_dir.name}] {len(todo)} new / {len(work)} total")
def _go(item):
try:
return runner(client, model, judge_model, item, n_frames)
except Exception as exc:
print(f"[{out_dir.name}] error on {item['video']}: {exc}", flush=True)
return None
n_done = 0
with ThreadPoolExecutor(max_workers=workers) as ex, open(results_path, "a") as out:
futures = {ex.submit(_go, item): item for item in todo}
for fut in as_completed(futures):
res = fut.result()
if res is None:
continue
out.write(json.dumps(res, ensure_ascii=False) + "\n")
out.flush()
n_done += 1
if n_done % 10 == 0 or n_done == len(todo):
print(f"[{out_dir.name}] {n_done}/{len(todo)} done", flush=True)
print(f"[{out_dir.name}] saved -> {results_path}")
# Compute and write metrics.json (matches existing eval_*.py schemas).
rows = [json.loads(l) for l in open(results_path) if l.strip()]
metrics = METRICS_FN[task_kind](rows, model, judge_model)
metrics_path = out_dir / "metrics.json"
with open(metrics_path, "w") as f:
json.dump(metrics, f, indent=2, ensure_ascii=False)
print(f"[{out_dir.name}] metrics -> {metrics_path}")
return results_path
# ---------------------------------------------------------------------------
# Main
# ---------------------------------------------------------------------------
def main():
ap = argparse.ArgumentParser()
ap.add_argument("--openai-key", default=os.environ.get("OPENAI_API_KEY"))
ap.add_argument("--model", default="gpt-5.5",
help="Model under test (frames-only multimodal).")
ap.add_argument("--judge-model", default="gpt-5.4")
ap.add_argument("--tasks", default="sync,mute,swap")
ap.add_argument("--workers", type=int, default=6)
ap.add_argument("--n-frames", type=int, default=8)
ap.add_argument("--label-suffix", default="_visualOnly",
help="Tagged into output dir name to flag the no-audio handicap.")
args = ap.parse_args()
if not args.openai_key:
sys.exit("[error] need --openai-key or OPENAI_API_KEY env")
client = _client(args.openai_key)
model_tag = args.model.replace("/", "_") + args.label_suffix
tasks = [t.strip() for t in args.tasks.split(",") if t.strip()]
if "sync" in tasks:
run_task(EVAL_ROOT / "sync" / f"sync_{model_tag}",
build_sync_work(), run_one_sync,
client, args.model, args.judge_model, args.n_frames, args.workers,
task_kind="sync")
if "mute" in tasks:
run_task(EVAL_ROOT / "mute" / f"mute_{model_tag}_promptNeutral",
build_mute_work(), run_one_mute,
client, args.model, args.judge_model, args.n_frames, args.workers,
task_kind="mute")
if "swap" in tasks:
run_task(EVAL_ROOT / "swap" / f"swap_{model_tag}_promptDirect",
build_swap_work(), run_one_swap,
client, args.model, args.judge_model, args.n_frames, args.workers,
task_kind="swap")
if __name__ == "__main__":
main()