Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -6,6 +6,7 @@ pipeline_tag: image-text-to-text
|
|
| 6 |
|
| 7 |

|
| 8 |
|
|
|
|
| 9 |
## Model Details
|
| 10 |
|
| 11 |
TRUST-VL is a unified and explainable vision-language model for general multimodal misinformation detection. It incorporates a novel Question-Aware Visual Amplifier module, designed to extract task-specific visual features. To support training, we also construct TRUST-Instruct, a large-scale instruction dataset containing 198K samples featuring structured reasoning chains aligned with human fact-checking workflows. Extensive experiments on both in-domain and zero-shot benchmarks demonstrate that TRUST-VL achieves state-of-the-art performance, while also offering strong generalization and interpretability.
|
|
|
|
| 6 |
|
| 7 |

|
| 8 |
|
| 9 |
+
|
| 10 |
## Model Details
|
| 11 |
|
| 12 |
TRUST-VL is a unified and explainable vision-language model for general multimodal misinformation detection. It incorporates a novel Question-Aware Visual Amplifier module, designed to extract task-specific visual features. To support training, we also construct TRUST-Instruct, a large-scale instruction dataset containing 198K samples featuring structured reasoning chains aligned with human fact-checking workflows. Extensive experiments on both in-domain and zero-shot benchmarks demonstrate that TRUST-VL achieves state-of-the-art performance, while also offering strong generalization and interpretability.
|