Distilling Human-Aligned Privacy Sensitivity Assessment from Large Language Models

This model is a lightweight encoder (150M parameters) based on BERT, designed for assessing the privacy sensitivity of textual data. It was introduced in the paper Distilling Human-Aligned Privacy Sensitivity Assessment from Large Language Models.

The model was distilled from Mistral Large 3 (675B) using a large-scale dataset of privacy-annotated texts across 10 diverse domains. It aims to preserve strong agreement with human judgments of privacy while being efficient enough for large-scale real-world deployment.

Model Details

How to Get Started with the Model

You can use the model directly with a Hugging Face pipeline:

from transformers import pipeline

classifier = pipeline("text-classification", model="gabrielloiseau/BERT-base-privacy")
result = classifier("Happy First Day of Spring!")
print(result)  # [{'label': '1', 'score': 0.98}]

The labels 1 through 5 represent the degree of privacy sensitivity, where 1 is the lowest and 5 is the highest.

Citation

@misc{loiseau2026distilling,
      title={Distilling Human-Aligned Privacy Sensitivity Assessment from Large Language Models}, 
      author={Gabriel Loiseau and Damien Sileo and Damien Riquet and Maxime Meyer and Marc Tommasi},
      year={2026},
      eprint={2603.29497},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2603.29497}, 
}
Downloads last month
25
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including gabrielloiseau/BERT-base-privacy

Paper for gabrielloiseau/BERT-base-privacy