Research access: usage conditions

This model is released for research purposes only. You must agree to the following conditions before accessing it.

This model is intended strictly for research. It must not be used in real-world systems, privacy-sensitive applications, or for processing or analyzing data about real individuals.

Log in or Sign Up to review the conditions and access this model content.

CLIP ViT-B/16 - SALMU Clean

Part of the SALMUBench benchmark for multimodal machine unlearning.

Paper: "SALMUBench: A Benchmark for Sensitive Association-Level Multimodal Unlearning" (CVPR 2026)

Project Page


How SALMUBench is used

Typical workflow:

  1. Start from the Compromised model
  2. Apply an unlearning method using the benchmark dataset
  3. Evaluate forgetting and utility metrics
  4. Compare results against the Clean reference model

Model description

This model is the reference model in SALMUBench.

It was trained without exposure to the SALMU sensitive associations, representing the behavior that a model should ideally have after successful unlearning.

Evaluation metrics compare the predictions of unlearned models against this clean reference.


Architecture and training

Architecture: CLIP ViT-B/16 (OpenCLIP implementation)

Training setup:

  • trained from scratch
  • ~400M image–text pairs
  • 32 training epochs
  • large-scale retain dataset derived from DataComp CommonPool
  • no SALMU sensitive data included

Related artifacts

Compromised model used for unlearning
clip-vit-b-16-salmu-compromised

SALMUBench evaluation dataset: salmubench-512-redistributed

SALMU training dataset (sensitive associations of the Compromised model): salmu-512-redistributed

Project repository: SALMUBench GitHub repository


Training data and usage

Training data

This model was trained on large-scale web-derived image–text data (DataComp), which may include real-world images of people. The original training data is not distributed by the authors.

Intended use

This model is intended for research on multimodal learning and machine unlearning, and for benchmarking within SALMUBench.

Limitations and risks

Like other large-scale models, it may encode biases or memorize patterns from training data, including associations involving real individuals. It is not designed for safety-critical applications.

Out-of-scope use

This model should not be used for biometric identification, surveillance, or profiling of individuals, or for inferring sensitive attributes from images.


Citation

@misc{selvassala2026salmubenchbenchmarksensitiveassociationlevel,
      title={SALMUBench: A Benchmark for Sensitive Association-Level Multimodal Unlearning}, 
      author={Cai Selvas-Sala and Lei Kang and Lluis Gomez},
      year={2026},
      eprint={2603.26316},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2603.26316}, 
}
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Paper for cvc-mmu/clip-vit-b-16-salmu-clean