openslr/librispeech_asr
Viewer • Updated • 585k • 100k • 222
How to use bgstud/whisper-small-libirClean-vs-commonNative-en with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("automatic-speech-recognition", model="bgstud/whisper-small-libirClean-vs-commonNative-en") # Load model directly
from transformers import AutoProcessor, AutoModelForSpeechSeq2Seq
processor = AutoProcessor.from_pretrained("bgstud/whisper-small-libirClean-vs-commonNative-en")
model = AutoModelForSpeechSeq2Seq.from_pretrained("bgstud/whisper-small-libirClean-vs-commonNative-en")# Load model directly
from transformers import AutoProcessor, AutoModelForSpeechSeq2Seq
processor = AutoProcessor.from_pretrained("bgstud/whisper-small-libirClean-vs-commonNative-en")
model = AutoModelForSpeechSeq2Seq.from_pretrained("bgstud/whisper-small-libirClean-vs-commonNative-en")This model is a fine-tuned version of openai/whisper-small on the librispeech_asr dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss | Wer |
|---|---|---|---|---|
| 1.2481 | 0.08 | 10 | 3.5688 | 21.1895 |
| 0.7793 | 0.16 | 20 | 2.8307 | 38.9990 |
| 0.5443 | 0.24 | 30 | 2.4196 | 67.0458 |
| 0.4484 | 0.32 | 40 | 2.2903 | 71.1732 |
| 0.4086 | 0.4 | 50 | 2.3358 | 85.5379 |
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("automatic-speech-recognition", model="bgstud/whisper-small-libirClean-vs-commonNative-en")