# Temporal Twins: A Matched-Control Benchmark for Temporal Fraud Detection Temporal Twins is a synthetic UPI-style temporal transaction benchmark where fraud and benign trajectories are statically matched but differ in delayed event-order structure. The benchmark is designed to test whether models can exploit temporal ordering under matched-prefix controls rather than relying on static transaction summaries or prefix-length shortcuts. ## Installation Recommended Python version: `3.11+` (`3.13` also works in the checked environment). ### pip ```bash pip install -r requirements.txt ``` ### conda If [environment.yml](environment.yml) is present: ```bash conda env create -f environment.yml conda activate temporal-twins ``` ## Repository Structure - `src/`: synthetic user, transaction, risk, fraud, graph, and core config code - `models/`: learned baselines and probe/oracle wrappers, including SeqGRU and temporal GNNs - `experiments/`: benchmark runner and matched-prefix evaluation code - `config/`: checked-in YAML configs used as base configs for experiments - `results/`: frozen experiment artifacts, including the final deterministic paper suite - `metadata/`: Croissant metadata and release-side validation notes - `release/`: manual-hosting bundle prepared for later upload ## Quick Smoke Test The public CLI supports a fast audit-mode smoke test: ```bash PYTHONPATH=. python3 experiments/run_all.py \ --fast \ --seed 0 \ --benchmark-mode temporal_twins_oracle_calib \ --experiments audit \ --device cpu ``` ## Exact Paper-Style Group Runner The checked-in CLI does **not** expose `--difficulty`, `--num-users`, or `--simulation-days` flags. The exact grouped reproductions below therefore use the existing helper functions in [experiments/run_all.py](experiments/run_all.py) through an inline Python wrapper. Define this shell helper once in your session: ```bash run_group() { local group="$1" local seed="$2" local out_json="$3" PYTHONPATH=. python3 - "$group" "$seed" "$out_json" <<'PY' import json import math import sys import time from pathlib import Path from src.core.config_loader import load_config from experiments.run_all import ( build_gate_pool_from_frames, gate_volume_is_sufficient, generate_single_difficulty, offset_gate_namespace, prepare_gate_subset, run_motif_validity_check, set_global_determinism, ) def normalize(value): if isinstance(value, dict): return {k: normalize(v) for k, v in value.items()} if isinstance(value, (list, tuple)): return [normalize(v) for v in value] if hasattr(value, "item"): try: value = value.item() except Exception: pass if isinstance(value, float) and not math.isfinite(value): return None return value group = sys.argv[1] seed = int(sys.argv[2]) out_json = Path(sys.argv[3]) if group == "oracle_calib": benchmark_mode = "temporal_twins_oracle_calib" difficulty = "easy" hard_abort = True force_temporal_models = True else: benchmark_mode = "temporal_twins" difficulty = group hard_abort = False force_temporal_models = True cfg = load_config("config/default.yaml") cfg = cfg.model_copy( update={ "num_users": 350, "simulation_days": 45, "benchmark_mode": benchmark_mode, "random_seed": seed, } ) set_global_determinism(seed) pool = generate_single_difficulty( cfg, difficulty=difficulty, seed=seed, benchmark_mode=benchmark_mode, ) gate = prepare_gate_subset(pool, seed=seed, fast_mode=False) pack_count = 1 while (not gate_volume_is_sufficient(gate["volume"], False)) and pack_count <= 6: extra_seed = seed + pack_count * 10007 extra_pack = generate_single_difficulty( cfg, difficulty=difficulty, seed=extra_seed, benchmark_mode=benchmark_mode, ) extra_pack = offset_gate_namespace(extra_pack, pack_count) pool = build_gate_pool_from_frames([pool, extra_pack]) gate = prepare_gate_subset(pool, seed=seed, fast_mode=False) pack_count += 1 gate["source_pool_events"] = int(len(pool)) gate["source_pool_pairs"] = int(pool.loc[pool["twin_pair_id"] >= 0, "twin_pair_id"].nunique()) if "twin_pair_id" in pool.columns else 0 gate["source_pool_packs"] = int(pack_count) start = time.time() gate_pass, report = run_motif_validity_check( df=pool, config=cfg, seed=seed, device="cpu", num_epochs=3, node_epochs=150, n_checkpoints=8, hard_abort=hard_abort, benchmark_mode=benchmark_mode, fast_mode=False, force_temporal_models=force_temporal_models, prebuilt_gate=gate, ) elapsed = time.time() - start result = { "benchmark_group": group, "benchmark_mode": benchmark_mode, "seed": seed, "primary_metric_label": report["audit_metric_label"], "secondary_metric_label": report["raw_metric_label"], "gate_pass": bool(gate_pass), "run_wall_time_sec": float(elapsed), **report, } out_json.parent.mkdir(parents=True, exist_ok=True) out_json.write_text(json.dumps(normalize(result), indent=2) + "\n") print(f"Wrote {out_json}") PY } ``` ## Reproduce Oracle Calibration Non-fast, reliable-volume `temporal_twins_oracle_calib`, seed `0`, `num_users=350`, `simulation_days=45`: ```bash run_group oracle_calib 0 results/paper_suite_repro/jobs/oracle_calib_0.json ``` ## Reproduce Easy / Medium / Hard Each command below reproduces the matched-prefix grouped benchmark for seed `0` with the paper-scale non-fast settings (`num_users=350`, `simulation_days=45`, `n_checkpoints=8`, deterministic CPU runtime): ```bash run_group easy 0 results/paper_suite_repro/jobs/easy_0.json run_group medium 0 results/paper_suite_repro/jobs/medium_0.json run_group hard 0 results/paper_suite_repro/jobs/hard_0.json ``` ## Reproduce Full Paper Suite There is no single checked-in `paper_suite` driver script. The exact grouped reproduction can be run as a shell loop over benchmark groups and seeds, followed by a small aggregation step that writes the artifact files: ### 1. Generate per-run JSON files ```bash mkdir -p results/paper_suite_repro/jobs for group in oracle_calib easy medium hard; do for seed in 0 1 2 3 4; do run_group "$group" "$seed" "results/paper_suite_repro/jobs/${group}_${seed}.json" done done ``` ### 2. Aggregate into paper-suite CSV and Markdown files ```bash PYTHONPATH=. python3 - <<'PY' import json from pathlib import Path import pandas as pd def summarize_mean_std(df, group_col): numeric_cols = [c for c in df.columns if c != group_col and pd.api.types.is_numeric_dtype(df[c])] grouped = df.groupby(group_col, dropna=False)[numeric_cols].agg(["mean", "std"]).reset_index() grouped.columns = [ group_col if col == group_col else f"{col}_{stat}" for col, stat in grouped.columns ] return grouped def volume_failures(row): fails = [] if row["matched_eval_pairs"] < 2000: fails.append(f"matched_eval_pairs={row['matched_eval_pairs']} (<2000)") if row["positives"] < 500: fails.append(f"positives={row['positives']} (<500)") if row["negatives"] < 500: fails.append(f"negatives={row['negatives']} (<500)") if row["unique_fraud_users"] < 50: fails.append(f"unique_fraud_users={row['unique_fraud_users']} (<50)") if row["unique_benign_users"] < 50: fails.append(f"unique_benign_users={row['unique_benign_users']} (<50)") if not (0.35 <= row["positive_rate"] <= 0.65): fails.append(f"positive_rate={row['positive_rate']:.4f} (outside [0.35,0.65])") return " | ".join(fails) def hard_gate_failures(row): checks = [ (row["primary_metric_label"], row["audit_roc_auc"], ">=", 0.99), (f"{row['primary_metric_label']} pair-sep", row["audit_pair_sep"], ">=", 0.99), (row["secondary_metric_label"], row["raw_roc_auc"], ">=", 0.95), (f"{row['secondary_metric_label']} pair-sep", row["raw_pair_sep"], ">=", 0.90), ("static_agg_auc", row["static_agg_auc"], "<=", 0.60), ("XGBoost ROC-AUC", row["xgb_roc_auc"], "<=", 0.65), ("StaticGNN ROC-AUC", row["static_gnn_roc"], "<=", 0.70), ("SeqGRU ROC-AUC", row["seqgru_roc_auc"], ">=", 0.80), ("SeqGRU shuffle delta", row["seqgru_shuffle_delta"], "<=", -0.10), ] fails = [] for label, value, op, threshold in checks: ok = value >= threshold if op == ">=" else value <= threshold if not ok: fails.append(f"{label}: {value:.4f} ({op}{threshold})") return " | ".join(fails) def advisory_failures(row): checks = [ ("TGN ROC-AUC", row["tgn_roc_auc"], ">=", 0.75), ("TGN shuffle delta", row["tgn_shuffle_delta"], "<=", -0.10), ("TGAT ROC-AUC", row["tgat_roc_auc"], ">=", 0.75), ("TGAT shuffle delta", row["tgat_shuffle_delta"], "<=", -0.10), ("DyRep ROC-AUC", row["dyrep_roc_auc"], ">=", 0.75), ("DyRep shuffle delta", row["dyrep_shuffle_delta"], "<=", -0.10), ("JODIE ROC-AUC", row["jodie_roc_auc"], ">=", 0.75), ("JODIE shuffle delta", row["jodie_shuffle_delta"], "<=", -0.10), ] fails = [] for label, value, op, threshold in checks: ok = value >= threshold if op == ">=" else value <= threshold if not ok: fails.append(f"{label}: {value:.4f} ({op}{threshold})") return " | ".join(fails) jobs_dir = Path("results/paper_suite_repro/jobs") out_dir = jobs_dir.parent rows = [json.loads(path.read_text()) for path in sorted(jobs_dir.glob("*.json"))] df = pd.DataFrame(rows).sort_values(["benchmark_group", "seed"]).reset_index(drop=True) runs_path = out_dir / "paper_suite_runs.csv" summary_path = out_dir / "paper_suite_summary.csv" runtime_path = out_dir / "paper_suite_runtime.csv" failed_path = out_dir / "paper_suite_failed_checks.csv" summary_md_path = out_dir / "paper_suite_summary.md" meta_path = out_dir / "paper_suite_meta.json" df.to_csv(runs_path, index=False) summary = summarize_mean_std(df, "benchmark_group") summary.to_csv(summary_path, index=False) runtime_cols = [ "benchmark_group", "seed", "run_wall_time_sec", "static_gnn_eval_time_sec", "static_gnn_unique_prefix_cutoffs", "static_gnn_graph_builds", "static_gnn_cache_hit_rate", ] df[runtime_cols].to_csv(runtime_path, index=False) failed = df[["benchmark_group", "seed", "gate_pass"]].copy() failed["volume_failures"] = df.apply(volume_failures, axis=1) failed["hard_gate_failures"] = df.apply(hard_gate_failures, axis=1) failed["advisory_failures"] = df.apply(advisory_failures, axis=1) failed.to_csv(failed_path, index=False) meta = { "device": "cpu", "num_users": 350, "simulation_days": 45, "num_epochs": 3, "node_epochs": 150, "n_checkpoints": 8, "fast_mode": False, "seeds": [0, 1, 2, 3, 4], } meta_path.write_text(json.dumps(meta, indent=2) + "\n") headline = summary[ [ "benchmark_group", "xgb_roc_auc_mean", "static_gnn_roc_mean", "seqgru_roc_auc_mean", "seqgru_shuffle_delta_mean", ] ].copy() lines = [ "# Paper Suite Summary", "", "| benchmark_group | xgb_roc_auc_mean | static_gnn_roc_mean | seqgru_roc_auc_mean | seqgru_shuffle_delta_mean |", "|---|---:|---:|---:|---:|", ] for row in headline.itertuples(index=False): lines.append( f"| {row.benchmark_group} | {row.xgb_roc_auc_mean:.4f} | {row.static_gnn_roc_mean:.4f} | {row.seqgru_roc_auc_mean:.4f} | {row.seqgru_shuffle_delta_mean:.4f} |" ) summary_md_path.write_text("\n".join(lines) + "\n") print(f"Wrote {runs_path}") print(f"Wrote {summary_path}") print(f"Wrote {runtime_path}") print(f"Wrote {failed_path}") print(f"Wrote {summary_md_path}") print(f"Wrote {meta_path}") PY ``` This aggregation step writes: - `results/paper_suite_repro/paper_suite_runs.csv` - `results/paper_suite_repro/paper_suite_summary.csv` - `results/paper_suite_repro/paper_suite_runtime.csv` - `results/paper_suite_repro/paper_suite_failed_checks.csv` - `results/paper_suite_repro/paper_suite_summary.md` - `results/paper_suite_repro/paper_suite_meta.json` The frozen reference artifacts checked into this repository live in [results/paper_suite_20260503_202810](results/paper_suite_20260503_202810). ## Expected Headline Results | benchmark_group | XGBoost ROC-AUC | StaticGNN ROC-AUC | SeqGRU ROC-AUC | SeqGRU shuffle delta | |---|---:|---:|---:|---:| | `oracle_calib` | 0.5000 | 0.5222 | 1.0000 | -0.5032 | | `easy` | 0.5000 | 0.4946 | 1.0000 | -0.5003 | | `medium` | 0.5000 | 0.4922 | 0.8391 | -0.3337 | | `hard` | 0.5000 | 0.5026 | 0.6876 | -0.1883 | ## Determinism - Deterministic CPU runtime is enabled in [experiments/run_all.py](experiments/run_all.py). - The same seed should produce identical matched-prefix data and identical metrics under the same deterministic environment. - Deterministic settings intentionally trade speed for repeatability and will slow larger runs. For more detail, see [docs/DETERMINISM.md](docs/DETERMINISM.md). ## Runtime Note Mean wall-clock runtime per benchmark group in the final deterministic paper suite: - `oracle_calib`: `1136.6s` - `easy`: `1345.9s` - `medium`: `2181.9s` - `hard`: `2613.7s` - cumulative summed runtime across all 20 runs: about `10.11` hours ## Data and Metadata - Dataset card: [DATASET_CARD.md](DATASET_CARD.md) - Croissant metadata: [metadata/temporal_twins_croissant.json](metadata/temporal_twins_croissant.json) - Manual-hosting release bundle: [release/](release/) Hosted URLs: - Dataset URL: [temporal-twins on Hugging Face](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins) - Code repository URL: [temporal-twins-code on Hugging Face](https://huggingface.co/temporal-twins-benchmark/temporal-twins-code) - Croissant metadata URL: [raw JSON-LD](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/raw/main/metadata/temporal_twins_croissant.json) - Croissant metadata browser page: [blob view](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/blob/main/metadata/temporal_twins_croissant.json) - Data URL: [data tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/data) - Results URL: [results tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/results) - Configs URL: [configs tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/configs) - Metadata URL: [metadata tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/metadata) - Paper or preprint: Not available during double-blind review; to be added after publication. ## License - Code: Apache License 2.0 (`Apache-2.0`) - Dataset and generated benchmark artifacts: Creative Commons Attribution 4.0 International (`CC-BY-4.0`) - Code SPDX-License-Identifier: `Apache-2.0` - Dataset SPDX-License-Identifier: `CC-BY-4.0` - No real UPI data or personal financial records are included. ## Citation Anonymous NeurIPS 2026 submission. The paper or preprint is not available during double-blind review and will be added after publication. ## Warning - Synthetic data only - No real UPI transactions - Not for production fraud deployment