Update README.md
Browse files
README.md
CHANGED
|
@@ -1,7 +1,6 @@
|
|
| 1 |
---
|
| 2 |
language:
|
| 3 |
- ro
|
| 4 |
-
license: cc-by-nc-4.0
|
| 5 |
library_name: pytorch
|
| 6 |
pipeline_tag: video-text-to-text
|
| 7 |
tags:
|
|
@@ -21,7 +20,7 @@ metrics:
|
|
| 21 |
|
| 22 |
This repository hosts the encoder-decoder VSR model checkpoints introduced in the paper *VSRo-200: A Romanian Visual Speech Recognition Dataset for Studying Supervision and Multimodal Robustness*.
|
| 23 |
|
| 24 |
-
The models are MultiVSR backbones fine-tuned on the **VSRo-200** corpus, a 200-hour collection of Romanian podcast recordings. For training code, data preparation scripts, and inference instructions, please refer to the [GitHub repository](https://
|
| 25 |
|
| 26 |
## Checkpoints
|
| 27 |
|
|
@@ -34,46 +33,79 @@ All checkpoints follow the naming pattern `model_[hours]_[type].pt`:
|
|
| 34 |
|
| 35 |
## Results
|
| 36 |
|
| 37 |
-
All results are reported in Word Error Rate (WER, %) on the **Test Unseen**
|
| 38 |
|
| 39 |
-
### Annotated
|
| 40 |
|
| 41 |
-
| Hours |
|
| 42 |
-
|:---:|:---:|:---:|
|
| 43 |
-
|
|
| 44 |
-
|
|
| 45 |
-
|
|
| 46 |
-
|
|
| 47 |
-
|
|
| 48 |
-
| 125 | — | 51.71 |
|
| 49 |
-
| 150 | — | 51.25 |
|
| 50 |
-
| 175 | — | 49.84 |
|
| 51 |
-
| 200 | — | 48.75 |
|
| 52 |
|
| 53 |
-
###
|
| 54 |
|
| 55 |
-
|
|
| 56 |
-
|:---|:---:|:---:|
|
| 57 |
-
|
|
| 58 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
|
| 60 |
### Gender bias analysis (40h models)
|
| 61 |
|
| 62 |
-
| Training subset | Global | Males | Females |
|
| 63 |
-
|:---|:---:|:---:|:---:|
|
| 64 |
-
| 40h Males | 62.15 | 61.32 | 62.97 |
|
| 65 |
-
| 40h Females | 59.33 | 59.17 | 59.49 |
|
| 66 |
-
| 40h Mix | 59.52 | 59.19 | 59.85 |
|
| 67 |
|
| 68 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 69 |
|
| 70 |
-
| OOD category | WER |
|
| 71 |
-
|:---|:---:|
|
| 72 |
-
| Vlogs | 58.61 |
|
| 73 |
-
| Specific domains | 63.01 |
|
| 74 |
-
| Noisy | 68.96 |
|
| 75 |
-
| Archival | 87.97 |
|
| 76 |
-
| Global OOD | 68.46 |
|
| 77 |
|
| 78 |
## Citation
|
| 79 |
|
|
@@ -85,4 +117,4 @@ If you use these models, please cite:
|
|
| 85 |
author = {...},
|
| 86 |
year = {...}
|
| 87 |
}
|
| 88 |
-
```
|
|
|
|
| 1 |
---
|
| 2 |
language:
|
| 3 |
- ro
|
|
|
|
| 4 |
library_name: pytorch
|
| 5 |
pipeline_tag: video-text-to-text
|
| 6 |
tags:
|
|
|
|
| 20 |
|
| 21 |
This repository hosts the encoder-decoder VSR model checkpoints introduced in the paper *VSRo-200: A Romanian Visual Speech Recognition Dataset for Studying Supervision and Multimodal Robustness*.
|
| 22 |
|
| 23 |
+
The models are MultiVSR backbones fine-tuned on the **VSRo-200** corpus, a 200-hour collection of Romanian podcast recordings. For training code, data preparation scripts, and inference instructions, please refer to the [GitHub repository](https://github.com/vsro200/vsro200).
|
| 24 |
|
| 25 |
## Checkpoints
|
| 26 |
|
|
|
|
| 33 |
|
| 34 |
## Results
|
| 35 |
|
| 36 |
+
All results are reported in Word Error Rate (WER, %) and Character Error Rate (CER, %) on the **Test Unseen** and **Test Seen** splits. Lower is better.
|
| 37 |
|
| 38 |
+
#### Human Annotated Data
|
| 39 |
|
| 40 |
+
| Training Hours | Test Unseen WER (%) | Test Unseen CER (%) | Test Seen WER (%) | Test Seen CER (%) |
|
| 41 |
+
|:---:|:---:|:---:|:---:|:---:|
|
| 42 |
+
| 10h | 72.50 | 41.49 | 67.01 | 37.53 |
|
| 43 |
+
| 25h | 64.86 | 36.62 | 59.23 | 32.96 |
|
| 44 |
+
| 50h | 58.87 | 33.38 | 54.03 | 29.88 |
|
| 45 |
+
| 75h | 54.86 | 30.97 | 51.44 | 28.61 |
|
| 46 |
+
| 100h | **53.29** | **29.94** | **48.16** | **26.53** |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
|
| 48 |
+
#### Whisper Pseudo Labels
|
| 49 |
|
| 50 |
+
| Training Hours | Test Unseen WER (%) | Test Unseen CER (%) | Test Seen WER (%) | Test Seen CER (%) |
|
| 51 |
+
|:---:|:---:|:---:|:---:|:---:|
|
| 52 |
+
| 10h | 74.61 | 42.09 | 68.41 | 38.22 |
|
| 53 |
+
| 25h | 66.27 | 37.05 | 60.40 | 33.36 |
|
| 54 |
+
| 50h | 59.28 | 33.15 | 55.39 | 30.65 |
|
| 55 |
+
| 75h | 56.25 | 31.18 | 51.56 | 28.33 |
|
| 56 |
+
| 100h | 53.63 | 30.12 | 49.61 | 27.22 |
|
| 57 |
+
| 125h | 51.71 | 29.04 | 48.68 | 26.58 |
|
| 58 |
+
| 150h | 51.25 | 28.40 | 47.05 | 25.64 |
|
| 59 |
+
| 175h | 49.84 | 27.66 | 46.44 | 25.30 |
|
| 60 |
+
| 200h | **48.75** | **27.05** | **44.54** | **24.51** |
|
| 61 |
+
|
| 62 |
+
A variance analysis across three random shuffles of the 100h subsets yields a mean Word Error Rate (WER) of 53.21% (± 0.37) for the human-annotated data and 53.82% (± 0.17) for the auto-generated data.
|
| 63 |
+
|
| 64 |
+
|
| 65 |
+
|
| 66 |
+
### Out-of-distribution robustness
|
| 67 |
+
|
| 68 |
+
* **Test Seen / Unseen (In-Domain):** Baseline performance on podcast data. "Seen" includes speakers present in the training set, while "Unseen" evaluates zero-shot speaker generalization.
|
| 69 |
+
* **Vlogs:** Unconstrained videos shot in varied, less controlled environments (different camera angles, dynamic lighting, movement).
|
| 70 |
+
* **Specific domains:** Content featuring highly specialized or technical vocabulary (e.g., medical, scientific). This category heavily tests the model's robustness to Out-Of-Vocabulary (OOV) words, exhibiting the highest OOV Type rate (17.93%).
|
| 71 |
+
* **Noisy:** Videos with poor resolution, bad lighting, or heavy motion blur.
|
| 72 |
+
* **Archival (Black & White):** Historical footage with distinct visual artifacts, atypical framerates, and lack of color information. This represents the hardest challenge for the visual front-end.
|
| 73 |
+
* **Global OOD:** The aggregated metrics across all out-of-distribution subsets, providing a single macro-score for the model's robustness in the wild.
|
| 74 |
+
|
| 75 |
+
| Dataset / Category | # Clips | WER (%) | CER (%) | OOV Token (%) | OOV Type (%) |
|
| 76 |
+
|:---|:---:|:---:|:---:|:---:|:---:|
|
| 77 |
+
| **Test Seen** | 386 | 44.54 | 24.51 | 1.67 | 6.93 |
|
| 78 |
+
| **Test Unseen** | 389 | 48.75 | 27.05 | 2.30 | 8.50 |
|
| 79 |
+
| **OOD: Vlogs** | 99 | 58.61 | 32.85 | 1.49 | 4.26 |
|
| 80 |
+
| **OOD: Specific domains** | 84 | 63.01 | 28.73 | 9.78 | 17.93 |
|
| 81 |
+
| **OOD: Noisy** | 100 | 68.96 | 33.68 | 6.19 | 12.88 |
|
| 82 |
+
| **OOD: Archival** | 92 | 87.97 | 50.44 | 5.24 | 10.96 |
|
| 83 |
+
| **Global OOD** | 375 | 68.46 | 35.99 | 5.08 | 14.75 |
|
| 84 |
+
|
| 85 |
+
#### Metrics Note
|
| 86 |
+
* **Duration:** Each OOD category consists of 15 minutes of video content.
|
| 87 |
+
* **OOV Token (%):** The percentage of *total words* in the evaluation set that do not appear in the training data. Measures how often unknown words occur.
|
| 88 |
+
* **OOV Type (%):** The percentage of *unique words* in the evaluation set that do not appear in the training data. Measures the diversity of unknown words.
|
| 89 |
|
| 90 |
### Gender bias analysis (40h models)
|
| 91 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
|
| 93 |
+
To evaluate gender bias and cross-speaker generalization, we trained 40-hour baseline models on male-only, female-only, and mixed datasets. The results reveal that while mixed data optimizes in-domain performance, training exclusively on female speakers provides more robust visual representations, leading to the best zero-shot generalization across both genders.
|
| 94 |
+
|
| 95 |
+
#### Test Unseen
|
| 96 |
+
| Training Set (40h) | Global WER (%) | Global CER (%) | Male WER (%) | Male CER (%) | Female WER (%) | Female CER (%) |
|
| 97 |
+
|:---|:---:|:---:|:---:|:---:|:---:|:---:|
|
| 98 |
+
| Males Only | 62.15 | 35.23 | 61.32 | 34.51 | 62.97 | 35.95 |
|
| 99 |
+
| Females Only | **59.33** | **33.44** | **59.17** | **32.87** | **59.49** | **34.02** |
|
| 100 |
+
| Mixed Data | 59.52 | 33.74 | 59.19 | 33.26 | 59.85 | 34.22 |
|
| 101 |
+
|
| 102 |
+
#### Test Seen
|
| 103 |
+
| Training Set (40h) | Global WER (%) | Global CER (%) | Male WER (%) | Male CER (%) | Female WER (%) | Female CER (%) |
|
| 104 |
+
|:---|:---:|:---:|:---:|:---:|:---:|:---:|
|
| 105 |
+
| Males Only | 58.82 | 33.11 | **58.58** | **32.59** | 59.06 | 33.63 |
|
| 106 |
+
| Females Only | 59.10 | 33.30 | 67.26 | 38.67 | **51.20** | **27.99** |
|
| 107 |
+
| Mixed Data | **56.29** | **31.22** | 60.56 | 33.54 | 52.15 | 28.93 |
|
| 108 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
|
| 110 |
## Citation
|
| 111 |
|
|
|
|
| 117 |
author = {...},
|
| 118 |
year = {...}
|
| 119 |
}
|
| 120 |
+
```
|