vsro200 commited on
Commit
7b13d18
·
verified ·
1 Parent(s): 3b1b958

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -34
README.md CHANGED
@@ -1,7 +1,6 @@
1
  ---
2
  language:
3
  - ro
4
- license: cc-by-nc-4.0
5
  library_name: pytorch
6
  pipeline_tag: video-text-to-text
7
  tags:
@@ -21,7 +20,7 @@ metrics:
21
 
22
  This repository hosts the encoder-decoder VSR model checkpoints introduced in the paper *VSRo-200: A Romanian Visual Speech Recognition Dataset for Studying Supervision and Multimodal Robustness*.
23
 
24
- The models are MultiVSR backbones fine-tuned on the **VSRo-200** corpus, a 200-hour collection of Romanian podcast recordings. For training code, data preparation scripts, and inference instructions, please refer to the [GitHub repository](https://https://github.com/vsro200/vsro200).
25
 
26
  ## Checkpoints
27
 
@@ -34,46 +33,79 @@ All checkpoints follow the naming pattern `model_[hours]_[type].pt`:
34
 
35
  ## Results
36
 
37
- All results are reported in Word Error Rate (WER, %) on the **Test Unseen** split. Lower is better.
38
 
39
- ### Annotated vs. auto-labeled data scaling
40
 
41
- | Hours | Annotated | Auto |
42
- |:---:|:---:|:---:|
43
- | 10 | 72.50 | 74.61 |
44
- | 25 | 64.86 | 66.27 |
45
- | 50 | 58.87 | 59.28 |
46
- | 75 | 54.86 | 56.25 |
47
- | 100 | 53.29 | 53.63 |
48
- | 125 | — | 51.71 |
49
- | 150 | — | 51.25 |
50
- | 175 | — | 49.84 |
51
- | 200 | — | 48.75 |
52
 
53
- ### Variance analysis (100h models, 3 random shuffles)
54
 
55
- | Data type | Mean WER | Std. dev. |
56
- |:---|:---:|:---:|
57
- | Human annotated | 53.21 | ± 0.37 |
58
- | Auto generated | 53.82 | ± 0.17 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59
 
60
  ### Gender bias analysis (40h models)
61
 
62
- | Training subset | Global | Males | Females |
63
- |:---|:---:|:---:|:---:|
64
- | 40h Males | 62.15 | 61.32 | 62.97 |
65
- | 40h Females | 59.33 | 59.17 | 59.49 |
66
- | 40h Mix | 59.52 | 59.19 | 59.85 |
67
 
68
- ### Out-of-distribution robustness (`model_200h_auto.pt`)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
 
70
- | OOD category | WER |
71
- |:---|:---:|
72
- | Vlogs | 58.61 |
73
- | Specific domains | 63.01 |
74
- | Noisy | 68.96 |
75
- | Archival | 87.97 |
76
- | Global OOD | 68.46 |
77
 
78
  ## Citation
79
 
@@ -85,4 +117,4 @@ If you use these models, please cite:
85
  author = {...},
86
  year = {...}
87
  }
88
- ```
 
1
  ---
2
  language:
3
  - ro
 
4
  library_name: pytorch
5
  pipeline_tag: video-text-to-text
6
  tags:
 
20
 
21
  This repository hosts the encoder-decoder VSR model checkpoints introduced in the paper *VSRo-200: A Romanian Visual Speech Recognition Dataset for Studying Supervision and Multimodal Robustness*.
22
 
23
+ The models are MultiVSR backbones fine-tuned on the **VSRo-200** corpus, a 200-hour collection of Romanian podcast recordings. For training code, data preparation scripts, and inference instructions, please refer to the [GitHub repository](https://github.com/vsro200/vsro200).
24
 
25
  ## Checkpoints
26
 
 
33
 
34
  ## Results
35
 
36
+ All results are reported in Word Error Rate (WER, %) and Character Error Rate (CER, %) on the **Test Unseen** and **Test Seen** splits. Lower is better.
37
 
38
+ #### Human Annotated Data
39
 
40
+ | Training Hours | Test Unseen WER (%) | Test Unseen CER (%) | Test Seen WER (%) | Test Seen CER (%) |
41
+ |:---:|:---:|:---:|:---:|:---:|
42
+ | 10h | 72.50 | 41.49 | 67.01 | 37.53 |
43
+ | 25h | 64.86 | 36.62 | 59.23 | 32.96 |
44
+ | 50h | 58.87 | 33.38 | 54.03 | 29.88 |
45
+ | 75h | 54.86 | 30.97 | 51.44 | 28.61 |
46
+ | 100h | **53.29** | **29.94** | **48.16** | **26.53** |
 
 
 
 
47
 
48
+ #### Whisper Pseudo Labels
49
 
50
+ | Training Hours | Test Unseen WER (%) | Test Unseen CER (%) | Test Seen WER (%) | Test Seen CER (%) |
51
+ |:---:|:---:|:---:|:---:|:---:|
52
+ | 10h | 74.61 | 42.09 | 68.41 | 38.22 |
53
+ | 25h | 66.27 | 37.05 | 60.40 | 33.36 |
54
+ | 50h | 59.28 | 33.15 | 55.39 | 30.65 |
55
+ | 75h | 56.25 | 31.18 | 51.56 | 28.33 |
56
+ | 100h | 53.63 | 30.12 | 49.61 | 27.22 |
57
+ | 125h | 51.71 | 29.04 | 48.68 | 26.58 |
58
+ | 150h | 51.25 | 28.40 | 47.05 | 25.64 |
59
+ | 175h | 49.84 | 27.66 | 46.44 | 25.30 |
60
+ | 200h | **48.75** | **27.05** | **44.54** | **24.51** |
61
+
62
+ A variance analysis across three random shuffles of the 100h subsets yields a mean Word Error Rate (WER) of 53.21% (± 0.37) for the human-annotated data and 53.82% (± 0.17) for the auto-generated data.
63
+
64
+
65
+
66
+ ### Out-of-distribution robustness
67
+
68
+ * **Test Seen / Unseen (In-Domain):** Baseline performance on podcast data. "Seen" includes speakers present in the training set, while "Unseen" evaluates zero-shot speaker generalization.
69
+ * **Vlogs:** Unconstrained videos shot in varied, less controlled environments (different camera angles, dynamic lighting, movement).
70
+ * **Specific domains:** Content featuring highly specialized or technical vocabulary (e.g., medical, scientific). This category heavily tests the model's robustness to Out-Of-Vocabulary (OOV) words, exhibiting the highest OOV Type rate (17.93%).
71
+ * **Noisy:** Videos with poor resolution, bad lighting, or heavy motion blur.
72
+ * **Archival (Black & White):** Historical footage with distinct visual artifacts, atypical framerates, and lack of color information. This represents the hardest challenge for the visual front-end.
73
+ * **Global OOD:** The aggregated metrics across all out-of-distribution subsets, providing a single macro-score for the model's robustness in the wild.
74
+
75
+ | Dataset / Category | # Clips | WER (%) | CER (%) | OOV Token (%) | OOV Type (%) |
76
+ |:---|:---:|:---:|:---:|:---:|:---:|
77
+ | **Test Seen** | 386 | 44.54 | 24.51 | 1.67 | 6.93 |
78
+ | **Test Unseen** | 389 | 48.75 | 27.05 | 2.30 | 8.50 |
79
+ | **OOD: Vlogs** | 99 | 58.61 | 32.85 | 1.49 | 4.26 |
80
+ | **OOD: Specific domains** | 84 | 63.01 | 28.73 | 9.78 | 17.93 |
81
+ | **OOD: Noisy** | 100 | 68.96 | 33.68 | 6.19 | 12.88 |
82
+ | **OOD: Archival** | 92 | 87.97 | 50.44 | 5.24 | 10.96 |
83
+ | **Global OOD** | 375 | 68.46 | 35.99 | 5.08 | 14.75 |
84
+
85
+ #### Metrics Note
86
+ * **Duration:** Each OOD category consists of 15 minutes of video content.
87
+ * **OOV Token (%):** The percentage of *total words* in the evaluation set that do not appear in the training data. Measures how often unknown words occur.
88
+ * **OOV Type (%):** The percentage of *unique words* in the evaluation set that do not appear in the training data. Measures the diversity of unknown words.
89
 
90
  ### Gender bias analysis (40h models)
91
 
 
 
 
 
 
92
 
93
+ To evaluate gender bias and cross-speaker generalization, we trained 40-hour baseline models on male-only, female-only, and mixed datasets. The results reveal that while mixed data optimizes in-domain performance, training exclusively on female speakers provides more robust visual representations, leading to the best zero-shot generalization across both genders.
94
+
95
+ #### Test Unseen
96
+ | Training Set (40h) | Global WER (%) | Global CER (%) | Male WER (%) | Male CER (%) | Female WER (%) | Female CER (%) |
97
+ |:---|:---:|:---:|:---:|:---:|:---:|:---:|
98
+ | Males Only | 62.15 | 35.23 | 61.32 | 34.51 | 62.97 | 35.95 |
99
+ | Females Only | **59.33** | **33.44** | **59.17** | **32.87** | **59.49** | **34.02** |
100
+ | Mixed Data | 59.52 | 33.74 | 59.19 | 33.26 | 59.85 | 34.22 |
101
+
102
+ #### Test Seen
103
+ | Training Set (40h) | Global WER (%) | Global CER (%) | Male WER (%) | Male CER (%) | Female WER (%) | Female CER (%) |
104
+ |:---|:---:|:---:|:---:|:---:|:---:|:---:|
105
+ | Males Only | 58.82 | 33.11 | **58.58** | **32.59** | 59.06 | 33.63 |
106
+ | Females Only | 59.10 | 33.30 | 67.26 | 38.67 | **51.20** | **27.99** |
107
+ | Mixed Data | **56.29** | **31.22** | 60.56 | 33.54 | 52.15 | 28.93 |
108
 
 
 
 
 
 
 
 
109
 
110
  ## Citation
111
 
 
117
  author = {...},
118
  year = {...}
119
  }
120
+ ```