NUSryan commited on
Commit
facde08
·
verified ·
1 Parent(s): 436ab4c

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -6,6 +6,7 @@ pipeline_tag: image-text-to-text
6
 
7
  ![trust-vl-logo](https://cdn-uploads.huggingface.co/production/uploads/65d01545ad23a67404c2d86d/DwNOl9uXkcLvlgB0Cbm29.png)
8
 
 
9
  ## Model Details
10
 
11
  TRUST-VL is a unified and explainable vision-language model for general multimodal misinformation detection. It incorporates a novel Question-Aware Visual Amplifier module, designed to extract task-specific visual features. To support training, we also construct TRUST-Instruct, a large-scale instruction dataset containing 198K samples featuring structured reasoning chains aligned with human fact-checking workflows. Extensive experiments on both in-domain and zero-shot benchmarks demonstrate that TRUST-VL achieves state-of-the-art performance, while also offering strong generalization and interpretability.
 
6
 
7
  ![trust-vl-logo](https://cdn-uploads.huggingface.co/production/uploads/65d01545ad23a67404c2d86d/DwNOl9uXkcLvlgB0Cbm29.png)
8
 
9
+
10
  ## Model Details
11
 
12
  TRUST-VL is a unified and explainable vision-language model for general multimodal misinformation detection. It incorporates a novel Question-Aware Visual Amplifier module, designed to extract task-specific visual features. To support training, we also construct TRUST-Instruct, a large-scale instruction dataset containing 198K samples featuring structured reasoning chains aligned with human fact-checking workflows. Extensive experiments on both in-domain and zero-shot benchmarks demonstrate that TRUST-VL achieves state-of-the-art performance, while also offering strong generalization and interpretability.