| # Temporal Twins: A Matched-Control Benchmark for Temporal Fraud Detection |
|
|
| Temporal Twins is a synthetic UPI-style temporal transaction benchmark where fraud and benign trajectories are statically matched but differ in delayed event-order structure. The benchmark is designed to test whether models can exploit temporal ordering under matched-prefix controls rather than relying on static transaction summaries or prefix-length shortcuts. |
|
|
| ## Installation |
|
|
| Recommended Python version: `3.11+` (`3.13` also works in the checked environment). |
|
|
| ### pip |
|
|
| ```bash |
| pip install -r requirements.txt |
| ``` |
|
|
| ### conda |
|
|
| If [environment.yml](environment.yml) is present: |
|
|
| ```bash |
| conda env create -f environment.yml |
| conda activate temporal-twins |
| ``` |
|
|
| ## Repository Structure |
|
|
| - `src/`: synthetic user, transaction, risk, fraud, graph, and core config code |
| - `models/`: learned baselines and probe/oracle wrappers, including SeqGRU and temporal GNNs |
| - `experiments/`: benchmark runner and matched-prefix evaluation code |
| - `config/`: checked-in YAML configs used as base configs for experiments |
| - `results/`: frozen experiment artifacts, including the final deterministic paper suite |
| - `metadata/`: Croissant metadata and release-side validation notes |
| - `release/`: manual-hosting bundle prepared for later upload |
|
|
| ## Quick Smoke Test |
|
|
| The public CLI supports a fast audit-mode smoke test: |
|
|
| ```bash |
| PYTHONPATH=. python3 experiments/run_all.py \ |
| --fast \ |
| --seed 0 \ |
| --benchmark-mode temporal_twins_oracle_calib \ |
| --experiments audit \ |
| --device cpu |
| ``` |
|
|
| ## Exact Paper-Style Group Runner |
|
|
| The checked-in CLI does **not** expose `--difficulty`, `--num-users`, or `--simulation-days` flags. The exact grouped reproductions below therefore use the existing helper functions in [experiments/run_all.py](experiments/run_all.py) through an inline Python wrapper. |
|
|
| Define this shell helper once in your session: |
|
|
| ```bash |
| run_group() { |
| local group="$1" |
| local seed="$2" |
| local out_json="$3" |
| |
| PYTHONPATH=. python3 - "$group" "$seed" "$out_json" <<'PY' |
| import json |
| import math |
| import sys |
| import time |
| from pathlib import Path |
| |
| from src.core.config_loader import load_config |
| from experiments.run_all import ( |
| build_gate_pool_from_frames, |
| gate_volume_is_sufficient, |
| generate_single_difficulty, |
| offset_gate_namespace, |
| prepare_gate_subset, |
| run_motif_validity_check, |
| set_global_determinism, |
| ) |
| |
| |
| def normalize(value): |
| if isinstance(value, dict): |
| return {k: normalize(v) for k, v in value.items()} |
| if isinstance(value, (list, tuple)): |
| return [normalize(v) for v in value] |
| if hasattr(value, "item"): |
| try: |
| value = value.item() |
| except Exception: |
| pass |
| if isinstance(value, float) and not math.isfinite(value): |
| return None |
| return value |
| |
| |
| group = sys.argv[1] |
| seed = int(sys.argv[2]) |
| out_json = Path(sys.argv[3]) |
| |
| if group == "oracle_calib": |
| benchmark_mode = "temporal_twins_oracle_calib" |
| difficulty = "easy" |
| hard_abort = True |
| force_temporal_models = True |
| else: |
| benchmark_mode = "temporal_twins" |
| difficulty = group |
| hard_abort = False |
| force_temporal_models = True |
| |
| cfg = load_config("config/default.yaml") |
| cfg = cfg.model_copy( |
| update={ |
| "num_users": 350, |
| "simulation_days": 45, |
| "benchmark_mode": benchmark_mode, |
| "random_seed": seed, |
| } |
| ) |
| |
| set_global_determinism(seed) |
| pool = generate_single_difficulty( |
| cfg, |
| difficulty=difficulty, |
| seed=seed, |
| benchmark_mode=benchmark_mode, |
| ) |
| gate = prepare_gate_subset(pool, seed=seed, fast_mode=False) |
| pack_count = 1 |
| |
| while (not gate_volume_is_sufficient(gate["volume"], False)) and pack_count <= 6: |
| extra_seed = seed + pack_count * 10007 |
| extra_pack = generate_single_difficulty( |
| cfg, |
| difficulty=difficulty, |
| seed=extra_seed, |
| benchmark_mode=benchmark_mode, |
| ) |
| extra_pack = offset_gate_namespace(extra_pack, pack_count) |
| pool = build_gate_pool_from_frames([pool, extra_pack]) |
| gate = prepare_gate_subset(pool, seed=seed, fast_mode=False) |
| pack_count += 1 |
| |
| gate["source_pool_events"] = int(len(pool)) |
| gate["source_pool_pairs"] = int(pool.loc[pool["twin_pair_id"] >= 0, "twin_pair_id"].nunique()) if "twin_pair_id" in pool.columns else 0 |
| gate["source_pool_packs"] = int(pack_count) |
| |
| start = time.time() |
| gate_pass, report = run_motif_validity_check( |
| df=pool, |
| config=cfg, |
| seed=seed, |
| device="cpu", |
| num_epochs=3, |
| node_epochs=150, |
| n_checkpoints=8, |
| hard_abort=hard_abort, |
| benchmark_mode=benchmark_mode, |
| fast_mode=False, |
| force_temporal_models=force_temporal_models, |
| prebuilt_gate=gate, |
| ) |
| elapsed = time.time() - start |
| |
| result = { |
| "benchmark_group": group, |
| "benchmark_mode": benchmark_mode, |
| "seed": seed, |
| "primary_metric_label": report["audit_metric_label"], |
| "secondary_metric_label": report["raw_metric_label"], |
| "gate_pass": bool(gate_pass), |
| "run_wall_time_sec": float(elapsed), |
| **report, |
| } |
| |
| out_json.parent.mkdir(parents=True, exist_ok=True) |
| out_json.write_text(json.dumps(normalize(result), indent=2) + "\n") |
| print(f"Wrote {out_json}") |
| PY |
| } |
| ``` |
|
|
| ## Reproduce Oracle Calibration |
|
|
| Non-fast, reliable-volume `temporal_twins_oracle_calib`, seed `0`, `num_users=350`, `simulation_days=45`: |
|
|
| ```bash |
| run_group oracle_calib 0 results/paper_suite_repro/jobs/oracle_calib_0.json |
| ``` |
|
|
| ## Reproduce Easy / Medium / Hard |
|
|
| Each command below reproduces the matched-prefix grouped benchmark for seed `0` with the paper-scale non-fast settings (`num_users=350`, `simulation_days=45`, `n_checkpoints=8`, deterministic CPU runtime): |
|
|
| ```bash |
| run_group easy 0 results/paper_suite_repro/jobs/easy_0.json |
| run_group medium 0 results/paper_suite_repro/jobs/medium_0.json |
| run_group hard 0 results/paper_suite_repro/jobs/hard_0.json |
| ``` |
|
|
| ## Reproduce Full Paper Suite |
|
|
| There is no single checked-in `paper_suite` driver script. The exact grouped reproduction can be run as a shell loop over benchmark groups and seeds, followed by a small aggregation step that writes the artifact files: |
|
|
| ### 1. Generate per-run JSON files |
|
|
| ```bash |
| mkdir -p results/paper_suite_repro/jobs |
| |
| for group in oracle_calib easy medium hard; do |
| for seed in 0 1 2 3 4; do |
| run_group "$group" "$seed" "results/paper_suite_repro/jobs/${group}_${seed}.json" |
| done |
| done |
| ``` |
|
|
| ### 2. Aggregate into paper-suite CSV and Markdown files |
|
|
| ```bash |
| PYTHONPATH=. python3 - <<'PY' |
| import json |
| from pathlib import Path |
| |
| import pandas as pd |
| |
| |
| def summarize_mean_std(df, group_col): |
| numeric_cols = [c for c in df.columns if c != group_col and pd.api.types.is_numeric_dtype(df[c])] |
| grouped = df.groupby(group_col, dropna=False)[numeric_cols].agg(["mean", "std"]).reset_index() |
| grouped.columns = [ |
| group_col if col == group_col else f"{col}_{stat}" |
| for col, stat in grouped.columns |
| ] |
| return grouped |
| |
| |
| def volume_failures(row): |
| fails = [] |
| if row["matched_eval_pairs"] < 2000: |
| fails.append(f"matched_eval_pairs={row['matched_eval_pairs']} (<2000)") |
| if row["positives"] < 500: |
| fails.append(f"positives={row['positives']} (<500)") |
| if row["negatives"] < 500: |
| fails.append(f"negatives={row['negatives']} (<500)") |
| if row["unique_fraud_users"] < 50: |
| fails.append(f"unique_fraud_users={row['unique_fraud_users']} (<50)") |
| if row["unique_benign_users"] < 50: |
| fails.append(f"unique_benign_users={row['unique_benign_users']} (<50)") |
| if not (0.35 <= row["positive_rate"] <= 0.65): |
| fails.append(f"positive_rate={row['positive_rate']:.4f} (outside [0.35,0.65])") |
| return " | ".join(fails) |
| |
| |
| def hard_gate_failures(row): |
| checks = [ |
| (row["primary_metric_label"], row["audit_roc_auc"], ">=", 0.99), |
| (f"{row['primary_metric_label']} pair-sep", row["audit_pair_sep"], ">=", 0.99), |
| (row["secondary_metric_label"], row["raw_roc_auc"], ">=", 0.95), |
| (f"{row['secondary_metric_label']} pair-sep", row["raw_pair_sep"], ">=", 0.90), |
| ("static_agg_auc", row["static_agg_auc"], "<=", 0.60), |
| ("XGBoost ROC-AUC", row["xgb_roc_auc"], "<=", 0.65), |
| ("StaticGNN ROC-AUC", row["static_gnn_roc"], "<=", 0.70), |
| ("SeqGRU ROC-AUC", row["seqgru_roc_auc"], ">=", 0.80), |
| ("SeqGRU shuffle delta", row["seqgru_shuffle_delta"], "<=", -0.10), |
| ] |
| fails = [] |
| for label, value, op, threshold in checks: |
| ok = value >= threshold if op == ">=" else value <= threshold |
| if not ok: |
| fails.append(f"{label}: {value:.4f} ({op}{threshold})") |
| return " | ".join(fails) |
| |
| |
| def advisory_failures(row): |
| checks = [ |
| ("TGN ROC-AUC", row["tgn_roc_auc"], ">=", 0.75), |
| ("TGN shuffle delta", row["tgn_shuffle_delta"], "<=", -0.10), |
| ("TGAT ROC-AUC", row["tgat_roc_auc"], ">=", 0.75), |
| ("TGAT shuffle delta", row["tgat_shuffle_delta"], "<=", -0.10), |
| ("DyRep ROC-AUC", row["dyrep_roc_auc"], ">=", 0.75), |
| ("DyRep shuffle delta", row["dyrep_shuffle_delta"], "<=", -0.10), |
| ("JODIE ROC-AUC", row["jodie_roc_auc"], ">=", 0.75), |
| ("JODIE shuffle delta", row["jodie_shuffle_delta"], "<=", -0.10), |
| ] |
| fails = [] |
| for label, value, op, threshold in checks: |
| ok = value >= threshold if op == ">=" else value <= threshold |
| if not ok: |
| fails.append(f"{label}: {value:.4f} ({op}{threshold})") |
| return " | ".join(fails) |
| |
| |
| jobs_dir = Path("results/paper_suite_repro/jobs") |
| out_dir = jobs_dir.parent |
| rows = [json.loads(path.read_text()) for path in sorted(jobs_dir.glob("*.json"))] |
| df = pd.DataFrame(rows).sort_values(["benchmark_group", "seed"]).reset_index(drop=True) |
| |
| runs_path = out_dir / "paper_suite_runs.csv" |
| summary_path = out_dir / "paper_suite_summary.csv" |
| runtime_path = out_dir / "paper_suite_runtime.csv" |
| failed_path = out_dir / "paper_suite_failed_checks.csv" |
| summary_md_path = out_dir / "paper_suite_summary.md" |
| meta_path = out_dir / "paper_suite_meta.json" |
| |
| df.to_csv(runs_path, index=False) |
| |
| summary = summarize_mean_std(df, "benchmark_group") |
| summary.to_csv(summary_path, index=False) |
| |
| runtime_cols = [ |
| "benchmark_group", |
| "seed", |
| "run_wall_time_sec", |
| "static_gnn_eval_time_sec", |
| "static_gnn_unique_prefix_cutoffs", |
| "static_gnn_graph_builds", |
| "static_gnn_cache_hit_rate", |
| ] |
| df[runtime_cols].to_csv(runtime_path, index=False) |
| |
| failed = df[["benchmark_group", "seed", "gate_pass"]].copy() |
| failed["volume_failures"] = df.apply(volume_failures, axis=1) |
| failed["hard_gate_failures"] = df.apply(hard_gate_failures, axis=1) |
| failed["advisory_failures"] = df.apply(advisory_failures, axis=1) |
| failed.to_csv(failed_path, index=False) |
| |
| meta = { |
| "device": "cpu", |
| "num_users": 350, |
| "simulation_days": 45, |
| "num_epochs": 3, |
| "node_epochs": 150, |
| "n_checkpoints": 8, |
| "fast_mode": False, |
| "seeds": [0, 1, 2, 3, 4], |
| } |
| meta_path.write_text(json.dumps(meta, indent=2) + "\n") |
| |
| headline = summary[ |
| [ |
| "benchmark_group", |
| "xgb_roc_auc_mean", |
| "static_gnn_roc_mean", |
| "seqgru_roc_auc_mean", |
| "seqgru_shuffle_delta_mean", |
| ] |
| ].copy() |
| |
| lines = [ |
| "# Paper Suite Summary", |
| "", |
| "| benchmark_group | xgb_roc_auc_mean | static_gnn_roc_mean | seqgru_roc_auc_mean | seqgru_shuffle_delta_mean |", |
| "|---|---:|---:|---:|---:|", |
| ] |
| for row in headline.itertuples(index=False): |
| lines.append( |
| f"| {row.benchmark_group} | {row.xgb_roc_auc_mean:.4f} | {row.static_gnn_roc_mean:.4f} | {row.seqgru_roc_auc_mean:.4f} | {row.seqgru_shuffle_delta_mean:.4f} |" |
| ) |
| summary_md_path.write_text("\n".join(lines) + "\n") |
| |
| print(f"Wrote {runs_path}") |
| print(f"Wrote {summary_path}") |
| print(f"Wrote {runtime_path}") |
| print(f"Wrote {failed_path}") |
| print(f"Wrote {summary_md_path}") |
| print(f"Wrote {meta_path}") |
| PY |
| ``` |
|
|
| This aggregation step writes: |
|
|
| - `results/paper_suite_repro/paper_suite_runs.csv` |
| - `results/paper_suite_repro/paper_suite_summary.csv` |
| - `results/paper_suite_repro/paper_suite_runtime.csv` |
| - `results/paper_suite_repro/paper_suite_failed_checks.csv` |
| - `results/paper_suite_repro/paper_suite_summary.md` |
| - `results/paper_suite_repro/paper_suite_meta.json` |
|
|
| The frozen reference artifacts checked into this repository live in [results/paper_suite_20260503_202810](results/paper_suite_20260503_202810). |
|
|
| ## Expected Headline Results |
|
|
| | benchmark_group | XGBoost ROC-AUC | StaticGNN ROC-AUC | SeqGRU ROC-AUC | SeqGRU shuffle delta | |
| |---|---:|---:|---:|---:| |
| | `oracle_calib` | 0.5000 | 0.5222 | 1.0000 | -0.5032 | |
| | `easy` | 0.5000 | 0.4946 | 1.0000 | -0.5003 | |
| | `medium` | 0.5000 | 0.4922 | 0.8391 | -0.3337 | |
| | `hard` | 0.5000 | 0.5026 | 0.6876 | -0.1883 | |
|
|
| ## Determinism |
|
|
| - Deterministic CPU runtime is enabled in [experiments/run_all.py](experiments/run_all.py). |
| - The same seed should produce identical matched-prefix data and identical metrics under the same deterministic environment. |
| - Deterministic settings intentionally trade speed for repeatability and will slow larger runs. |
|
|
| For more detail, see [docs/DETERMINISM.md](docs/DETERMINISM.md). |
|
|
| ## Runtime Note |
|
|
| Mean wall-clock runtime per benchmark group in the final deterministic paper suite: |
|
|
| - `oracle_calib`: `1136.6s` |
| - `easy`: `1345.9s` |
| - `medium`: `2181.9s` |
| - `hard`: `2613.7s` |
| - cumulative summed runtime across all 20 runs: about `10.11` hours |
|
|
| ## Data and Metadata |
|
|
| - Dataset card: [DATASET_CARD.md](DATASET_CARD.md) |
| - Croissant metadata: [metadata/temporal_twins_croissant.json](metadata/temporal_twins_croissant.json) |
| - Manual-hosting release bundle: [release/](release/) |
|
|
| Hosted URLs: |
|
|
| - Dataset URL: [temporal-twins on Hugging Face](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins) |
| - Code repository URL: [temporal-twins-code on Hugging Face](https://huggingface.co/temporal-twins-benchmark/temporal-twins-code) |
| - Croissant metadata URL: [raw JSON-LD](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/raw/main/metadata/temporal_twins_croissant.json) |
| - Croissant metadata browser page: [blob view](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/blob/main/metadata/temporal_twins_croissant.json) |
| - Data URL: [data tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/data) |
| - Results URL: [results tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/results) |
| - Configs URL: [configs tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/configs) |
| - Metadata URL: [metadata tree](https://huggingface.co/datasets/temporal-twins-benchmark/temporal-twins/tree/main/metadata) |
| - Paper or preprint: Not available during double-blind review; to be added after publication. |
|
|
| ## License |
|
|
| - Code: Apache License 2.0 (`Apache-2.0`) |
| - Dataset and generated benchmark artifacts: Creative Commons Attribution 4.0 International (`CC-BY-4.0`) |
| - Code SPDX-License-Identifier: `Apache-2.0` |
| - Dataset SPDX-License-Identifier: `CC-BY-4.0` |
| - No real UPI data or personal financial records are included. |
|
|
| ## Citation |
|
|
| Anonymous NeurIPS 2026 submission. The paper or preprint is not available during double-blind review and will be added after publication. |
|
|
| ## Warning |
|
|
| - Synthetic data only |
| - No real UPI transactions |
| - Not for production fraud deployment |
|
|