| """Re-run Gemini mute on the 6 case-study candidates with NEUTRAL prompt. |
| |
| Direct prompt failures for Gemini all turned out to be empty responses (API |
| artifacts), not real audio hallucinations. Neutral prompt ("Describe the audio |
| you hear") forces an open-ended answer; OpenAI judge classifies whether the |
| response hallucinates audio content vs correctly reports silence. |
| |
| Pipeline: |
| 1. Read /home/ubuntu/case_study_candidates.jsonl (6 base videos). |
| 2. Build a 6-row test jsonl (mute loader dedupes to base originals only). |
| 3. Run scripts/eval_gemini_mute_sync_swap.py --tasks mute |
| --mute-prompt-mode neutral --openai-judge. |
| 4. Print per-video raw_output + judge classification. |
| |
| Usage: |
| GEMINI_API_KEY=... OPENAI_API_KEY=... python3 /home/ubuntu/case_study_gemini_mute_neutral.py |
| """ |
|
|
| import argparse |
| import json |
| import os |
| import subprocess |
| import sys |
| from pathlib import Path |
|
|
| REPO = Path("/home/ubuntu/CleverHans-Evaluation") |
| TEST_JSONL = REPO / "data" / "kto_training_data_v2_test.jsonl" |
| EVAL_SCRIPT = REPO / "scripts" / "eval_gemini_mute_sync_swap.py" |
| DATA_ROOT = Path("/opt/dlami/nvme/video_source") |
| CANDIDATES = Path("/home/ubuntu/case_study_pool.jsonl") |
|
|
| OUT_ROOT = Path("/home/ubuntu/eval_results/case_study_pool") |
| FILTERED_JSON = OUT_ROOT / "mute_test.jsonl" |
| LABEL = "case_study_pool_gemini_mute_neutral" |
| GEMINI_RESULT = OUT_ROOT / "mute" / LABEL / "eval_results.jsonl" |
|
|
|
|
| def load_jsonl(path): |
| with open(path) as f: |
| for line in f: |
| line = line.strip() |
| if line: |
| yield json.loads(line) |
|
|
|
|
| def step_filter(): |
| bases = {row["video"] for row in load_jsonl(CANDIDATES)} |
| print(f"[filter] {len(bases)} base videos from {CANDIDATES}") |
|
|
| OUT_ROOT.mkdir(parents=True, exist_ok=True) |
| kept = [] |
| for row in load_jsonl(TEST_JSONL): |
| v = row["video"] |
| if "_delay_" in v or "_early_" in v: |
| continue |
| if v in bases: |
| kept.append(row) |
|
|
| with open(FILTERED_JSON, "w") as f: |
| for row in kept: |
| f.write(json.dumps(row, ensure_ascii=False) + "\n") |
| print(f"[filter] kept {len(kept)} rows -> {FILTERED_JSON}") |
| if len(kept) != len(bases): |
| print(f"[warn] expected {len(bases)} rows, got {len(kept)}") |
|
|
|
|
| def step_run(gemini_key, openai_key, model, workers): |
| env = os.environ.copy() |
| env["GEMINI_API_KEY"] = gemini_key |
| env["OPENAI_API_KEY"] = openai_key |
|
|
| cmd = [ |
| sys.executable, str(EVAL_SCRIPT), |
| "--tasks", "mute", |
| "--model", model, |
| "--data-root", str(DATA_ROOT), |
| "--test-jsonl", str(FILTERED_JSON), |
| "--output-dir", str(OUT_ROOT), |
| "--label", LABEL, |
| "--mute-prompt-mode", "neutral", |
| "--openai-judge", |
| "--workers", str(workers), |
| ] |
| print(f"[run] {' '.join(cmd)}") |
| subprocess.run(cmd, env=env, check=True) |
|
|
|
|
| def step_report(): |
| if not GEMINI_RESULT.exists(): |
| sys.exit(f"[error] expected results at {GEMINI_RESULT}") |
| rows = list(load_jsonl(GEMINI_RESULT)) |
| print(f"\n[report] {len(rows)} mute samples (neutral, OpenAI judged):") |
| for r in rows: |
| v = r["video"] |
| pred = r.get("pred_label") |
| correct = r.get("correct") |
| raw = (r.get("raw_output") or "").strip().replace("\n", " ")[:240] |
| je = (r.get("judge_explanation") or "").strip()[:120] |
| print(f"\n VIDEO: {v}") |
| print(f" pred={pred} correct={correct}") |
| print(f" raw : {raw!r}") |
| if je: |
| print(f" judge: {je}") |
|
|
|
|
| def main(): |
| p = argparse.ArgumentParser() |
| p.add_argument("--gemini-key", default=os.environ.get("GEMINI_API_KEY")) |
| p.add_argument("--openai-key", default=os.environ.get("OPENAI_API_KEY")) |
| p.add_argument("--model", default="gemini-3.1-pro-preview") |
| p.add_argument("--workers", type=int, default=4) |
| p.add_argument("--skip-run", action="store_true") |
| args = p.parse_args() |
|
|
| step_filter() |
| if not args.skip_run: |
| if not args.gemini_key: |
| sys.exit("[error] need --gemini-key or GEMINI_API_KEY env") |
| if not args.openai_key: |
| sys.exit("[error] need --openai-key or OPENAI_API_KEY env") |
| step_run(args.gemini_key, args.openai_key, args.model, args.workers) |
| step_report() |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|