Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
cpu
float64
31.1
58.9
memory
float64
53.2
71.4
disk_io
float64
74.7
162
network
float64
29.3
71.9
errors
int64
0
3
label
int64
0
0
47.48
61.59
129.72
62.18
0
0
46.4
65.03
111.29
45.8
0
0
46.21
56.26
94.13
45.5
1
0
40.46
57.76
141.98
48.19
0
0
57.8
63.18
121.83
45.88
0
0
42
61.12
110.97
64.82
0
0
42.33
61.98
116.56
53.11
1
0
58.89
65.58
123.28
57.05
1
0
39.95
57.25
131.61
45.69
0
0
46.72
56.71
124.86
46.92
0
0
47.91
64.66
133.41
56.04
0
0
40.8
61.07
124.97
57.8
0
0
44.11
60.77
137.7
42.81
1
0
39.81
61.77
134.59
56.36
0
0
44.82
66.69
80.7
56.58
0
0
37.91
65.13
133.55
50.15
0
0
52.39
60.45
107.87
45.99
2
0
46.29
62.94
140.58
51.4
1
0
43.72
60.9
139.11
47.66
1
0
31.72
63.04
114.07
47.69
0
0
44.19
63.21
148.29
51.4
0
0
44.91
56.98
103.91
42.06
0
0
45.3
69.39
117.11
52.41
0
0
46.7
61.98
131.51
40.8
1
0
40.99
66.15
141.08
61.14
0
0
40.05
60.3
121.49
45.97
1
0
37.25
62.21
104.07
53.79
1
0
41.08
61.03
132.2
40.15
1
0
36.96
62.55
123.9
56.25
0
0
51.5
59.4
129.26
59.74
0
0
46.47
59.86
147.99
53.79
1
0
40.13
64.36
137.38
43.43
1
0
49.11
67.69
116.32
43.97
0
0
49.39
61.28
138.14
54.3
0
0
43.68
70.16
129.39
43.14
0
0
45.21
61.59
118.14
55.92
3
0
50.23
60.97
106.11
45.9
0
0
45.87
63.16
106.74
51.23
2
0
50.29
62.15
133.09
58.53
1
0
49.53
60.19
124.57
52.06
1
0
52.19
62.02
139.96
57.91
2
0
41.14
61.29
112.72
50.65
0
0
36.06
63.07
113.78
50.64
0
0
42.64
65.27
120.96
41.38
1
0
41.35
62.65
120.68
44.79
0
0
51.82
64.49
120.23
53.54
0
0
41.69
64.56
108.11
49.08
0
0
41.14
66.98
114.99
47.46
0
0
53.83
63.21
101.09
57.34
2
0
38.37
67.53
109.04
39.76
1
0
43.81
67.62
119.71
42.54
0
0
43.74
58.26
144.49
38.56
1
0
52.21
57.69
137.45
50.08
1
0
46
60.2
121.05
46.92
1
0
52.93
58.29
152
34.38
1
0
46.4
60.13
116.88
46.06
1
0
42.05
64.55
125.36
44.46
1
0
49.06
63.89
107.57
45.52
1
0
44.9
62.35
139.16
45.27
0
0
52.76
60.68
113.88
52.09
1
0
42.38
64.85
116.41
56.83
0
0
46.62
61.61
121.45
54.76
1
0
39.97
58.36
137.37
56.33
2
0
41.5
61.97
88.65
64.16
1
0
48.07
66.07
113.5
51.63
1
0
50.77
60.57
110.04
60.88
0
0
46.55
66.43
132.86
48.72
0
0
43.49
61.99
109.76
49.61
0
0
46.61
59.52
127.79
62.26
1
0
48.45
60.8
123.36
50.1
0
0
47.37
62.16
154.14
53.25
1
0
51
67.45
118.34
57.81
1
0
44.21
57.93
135.49
35.52
0
0
46.71
67.63
134.26
45.38
1
0
38.4
67.49
137.69
46.25
1
0
36.43
66.06
118.28
59.9
0
0
44.46
61.83
123.89
52.42
2
0
42.75
63.87
103.99
48.86
1
0
48.56
58.63
96.99
60.22
0
0
46.21
68.14
121.72
62.41
1
0
50.13
62.5
104.69
47.85
0
0
48.23
66.11
105.53
55.49
0
0
36.35
66.33
133.29
48.9
0
0
52.51
62.22
144.43
38.96
2
0
44.93
64.39
115.78
36.68
1
0
50.61
57.68
118.16
49.4
0
0
46.83
59.18
112.29
41.53
1
0
40.07
63.51
112.05
43.66
0
0
37.83
61.43
120.86
31.74
1
0
47.57
65.61
132.77
45.04
1
0
42.26
58.32
155.61
47.51
0
0
53.22
61.25
128.65
52.49
0
0
40.14
62.24
142.86
30.35
0
0
41.22
57.73
110.3
41.35
0
0
46.09
64.73
135.83
55.23
0
0
45.39
59.42
142.85
54.31
0
0
43.06
59.99
140.94
39.52
0
0
38.01
63.69
110.24
46.1
0
0
40.86
60.36
82.46
55.48
1
0
46.95
61.75
123.56
52.19
0
0
End of preview. Expand in Data Studio

WaveGuard Anomaly Detection Benchmarks

Curated benchmark datasets and comparison results for evaluating anomaly detection models. Includes labeled training (normal) and test (mixed normal + anomalous) splits, plus head-to-head comparisons between WaveGuard and traditional methods.

Benchmark Comparisons (benchmark_results/)

WaveGuard vs. IsolationForest, LOF, and OneClassSVM across 12 datasets.

Summary: WaveGuard ranked #1 on all 12 datasets by F1 score.

Dataset WaveGuard IsolationForest LOF OneClassSVM Winner
Credit Card Fraud* 0.653 0.607 0.601 0.472 WaveGuard
Network Intrusion* 0.598 0.252 0.232 0.546 WaveGuard
Crypto Fraud 1.000 0.933 0.946 0.897 WaveGuard
Prompt Injection 0.976 0.952 0.976 0.889 WaveGuard
Phish Guard 0.976 0.905 0.952 0.816 WaveGuard
Content Guard 0.975 0.842 0.879 0.784 WaveGuard
Fraud Lens 0.949 0.896 0.882 0.800 WaveGuard
Ad Click Fraud 0.988 0.952 0.930 0.889 WaveGuard
Insurance Claims 0.972 0.921 0.959 0.833 WaveGuard
Network Security 0.990 0.962 0.980 0.952 WaveGuard
API Monitoring 0.959 0.909 0.933 0.814 WaveGuard
Log Anomalies 0.946 0.875 0.875 0.805 WaveGuard

*Real-world datasets. Others use domain-specific test suites with realistic feature schemas.

See benchmark_results/comparison.json for full details including sample sizes, feature counts, and anomaly rates.

Datasets

1. Server Metrics (server_metrics/)

Simulated server health metrics with injected failure events.

  • Features: cpu, memory, disk_io, network, errors (5 numeric)
  • Training: 500 normal samples
  • Test: 100 samples (15 anomalous)
  • Anomaly types: CPU spike, memory leak, disk saturation, network flood

2. Synthetic Time Series (synthetic_timeseries/)

Controlled synthetic signals with known anomaly injection points.

  • Patterns: sinusoidal, trend, seasonal, random walk
  • Anomaly types: point (spike), contextual (subtle shift), collective (regime change)
  • Training: 200 clean windows per pattern
  • Test: 50 windows per pattern (10 anomalous each)

Format

Each dataset is provided as Parquet files:

dataset_name/
  train.parquet     # Normal samples only
  test.parquet      # Mixed normal + anomalous
  metadata.json     # Feature descriptions, anomaly counts, creation params

Usage

from datasets import load_dataset

ds = load_dataset("emergentphysicslab/waveguard-benchmarks", "server_metrics")
train = ds["train"].to_pandas()
test = ds["test"].to_pandas()

Evaluation Protocol

  1. Train/fit your detector on train.parquet only
  2. Score each row in test.parquet
  3. Report: Precision, Recall, F1, AUC-ROC, Average Latency
  4. Compare against WaveGuard baseline in the model card

Citation

@dataset{waveguard_benchmarks2026,
  title={WaveGuard Anomaly Detection Benchmarks},
  author={Partin, Greg},
  year={2026},
  url={https://huggingface.co/datasets/emergentphysicslab/waveguard-benchmarks}
}
Downloads last month
49

Models trained or fine-tuned on emergentphysicslab/waveguard-benchmarks

Space using emergentphysicslab/waveguard-benchmarks 1