neurips26 commited on
Commit
9f8d2a5
Β·
verified Β·
1 Parent(s): 1c5c815

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - text-generation
5
+ - image-text-to-text
6
+ language:
7
+ - en
8
+ tags:
9
+ - machine-unlearning
10
+ - multimodal
11
+ - benchmark
12
+ - evaluation
13
+ - privacy
14
+ - llm
15
+ - vlm
16
+ - neurips
17
+ pretty_name: Multimodal Unlearning Evaluation Benchmark
18
+ size_categories:
19
+ - n<1K
20
+ ---
21
+ # 🧠 Multimodal Unlearning Evaluation Benchmark
22
+
23
+ ## πŸ“Œ Overview
24
+ This dataset provides evaluation outputs for studying **metric inconsistency in multimodal machine unlearning**.
25
+
26
+ It supports reproducibility of results in:
27
+ > *Metric Unreliability in Multimodal Machine Unlearning (NeurIPS 2026)*
28
+
29
+ ---
30
+
31
+ ## πŸ“Š Contents
32
+
33
+ | File | Description |
34
+ |------|------------|
35
+ | πŸ“„ `multimodal_results.json` | Results on VQA benchmarks (MLLMU, UnLOK, MMUBench) |
36
+ | πŸ“„ `unimodal_results.json` | CIFAR-10 baseline results |
37
+ | βš–οΈ `uqs_weights.json` | Learned weights for Unified Quality Score (UQS) |
38
+ | πŸ† `ranking_table.json` | Method rankings across metrics |
39
+ | πŸ“ˆ `analysis_results.json` | Correlation + disagreement analysis |
40
+ | πŸ” `kr_pilot_results.json` | Knowledge Recoverability (KR) pilot results |
41
+ | πŸ€– `blip2_minimal_summary.json` | Cross-architecture validation (BLIP-2) |
42
+
43
+ ---
44
+
45
+ ## 🎯 Purpose
46
+
47
+ This benchmark evaluates five standard unlearning metrics:
48
+
49
+ - Forget Accuracy (FA)
50
+ - Retain Accuracy (RA)
51
+ - Membership Inference Attack (MIA)
52
+ - Activation Distance (AD)
53
+ - JS Divergence (JS)
54
+
55
+ ⚠️ Key finding:
56
+ > These metrics produce **conflicting rankings** and do not measure **knowledge recoverability (KR)**.
57
+
58
+ ---
59
+
60
+ ## βš™οΈ Usage
61
+
62
+ All results in the paper can be reproduced directly from these files.
63
+
64
+ Example:
65
+
66
+ ```python
67
+ import json
68
+
69
+ with open("multimodal_results.json") as f:
70
+ data = json.load(f)
71
+
72
+
73
+ πŸ“š Source Datasets (Not Included)
74
+
75
+ This benchmark builds on:
76
+
77
+ MLLMU-Bench
78
+ UnLOK-VQA
79
+ MMUBench
80
+ CIFAR-10
81
+
82
+ These datasets are not redistributed here. Please refer to their original sources.
83
+
84
+ βš–οΈ License
85
+
86
+ This dataset is released under the CC-BY-4.0 License.
87
+
88
+ ⚠️ Notes
89
+ This dataset contains evaluation outputs, not raw training data
90
+ Designed for benchmarking and reproducibility
91
+ Prepared to support anonymous peer review
92
+ πŸ”— Citation
93
+ Anonymous. Metric Unreliability in Multimodal Machine Unlearning. NeurIPS 2026.
94
+
95
+ ---