Whisper Large v3 - Japanese Zatoichi ASR

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4736
  • Wer: 66.5438
  • Cer: 19.8215

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1.4e-05
  • train_batch_size: 32
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.0467 0.0164 1 1.2545 98.1567 40.8419
1.1366 0.0328 2 0.9981 100.8295 40.1226
0.9628 0.0492 3 0.9694 94.1935 37.4983
0.8432 0.0656 4 0.9222 89.9539 35.2604
0.8449 0.0820 5 0.8580 88.9401 33.0358
0.7821 0.0984 6 0.8142 90.8756 32.3565
0.8153 0.1148 7 0.7644 90.7834 36.8856
0.8144 0.1311 8 0.7391 93.4562 38.6306
0.6041 0.1475 9 0.7100 95.9447 37.8447
0.564 0.1639 10 0.6911 100.9217 36.4859
0.7197 0.1803 11 0.6618 100.0 37.7514
0.6799 0.1967 12 0.6438 107.0046 37.1920
0.7038 0.2131 13 0.6333 90.1382 31.2242
0.6186 0.2295 14 0.6190 83.1336 27.7608
0.5577 0.2459 15 0.6090 84.2396 27.7608
0.5531 0.2623 16 0.6070 76.7742 24.5904
0.4753 0.2787 17 0.6016 82.5806 26.7084
0.4771 0.2951 18 0.5959 83.5945 30.4249
0.5545 0.3115 19 0.5877 79.3548 25.4562
0.4436 0.3279 20 0.5851 82.3963 28.4268
0.5181 0.3443 21 0.5785 81.6590 27.9606
0.5177 0.3607 22 0.5716 78.0645 26.6818
0.5366 0.3770 23 0.5796 77.4194 28.8531
0.5237 0.3934 24 0.5762 74.0092 26.2155
0.5373 0.4098 25 0.5669 72.3502 23.6180
0.4487 0.4262 26 0.5599 77.3272 24.7236
0.395 0.4426 27 0.5577 73.3641 23.3249
0.4817 0.4590 28 0.5542 78.8940 24.9900
0.4511 0.4754 29 0.5588 78.3410 24.8435
0.4358 0.4918 30 0.5522 72.7189 23.4448
0.5482 0.5082 31 0.5481 72.4424 23.5380
0.5053 0.5246 32 0.5495 71.7051 23.1917
0.4177 0.5410 33 0.5474 72.8111 23.1384
0.5667 0.5574 34 0.5406 73.4562 23.0318
0.4687 0.5738 35 0.5354 73.7327 23.1917
0.4541 0.5902 36 0.5319 80.4608 25.7360
0.4454 0.6066 37 0.5270 79.5392 25.4829
0.511 0.6230 38 0.5208 73.4562 22.6056
0.4703 0.6393 39 0.5167 71.5207 22.4191
0.5256 0.6557 40 0.5173 70.5069 21.9662
0.4419 0.6721 41 0.5152 71.8894 21.9928
0.4823 0.6885 42 0.5109 73.5484 24.5238
0.4552 0.7049 43 0.5058 74.1935 24.9234
0.4432 0.7213 44 0.5016 70.0461 22.2326
0.3936 0.7377 45 0.4983 69.4931 21.5799
0.4348 0.7541 46 0.4972 71.1521 21.9395
0.4879 0.7705 47 0.4965 70.5991 22.2193
0.4758 0.7869 48 0.4964 70.1382 22.0594
0.3998 0.8033 49 0.4966 69.6774 22.2059
0.5079 0.8197 50 0.4973 70.1382 22.0328
0.4743 0.8361 51 0.4961 70.5991 22.0994
0.3918 0.8525 52 0.4937 69.4931 21.6465
0.5702 0.8689 53 0.4924 69.4931 21.7131
0.345 0.8852 54 0.4903 73.3641 23.2849
0.4643 0.9016 55 0.4890 73.1797 23.2183
0.4393 0.9180 56 0.4905 73.0876 23.2716
0.3611 0.9344 57 0.4934 72.4424 23.2450
0.4097 0.9508 58 0.4924 71.7972 23.6180
0.4346 0.9672 59 0.4885 72.6267 23.6446
0.4457 0.9836 60 0.4852 72.9032 24.6170
0.3451 1.0 61 0.4849 72.6267 23.4448
0.3444 1.0164 62 0.4854 73.2719 23.0718
0.2574 1.0328 63 0.4859 68.3871 21.4866
0.1913 1.0492 64 0.4868 69.0323 21.6065
0.2677 1.0656 65 0.4882 67.1889 21.7397
0.2288 1.0820 66 0.4890 68.1106 21.1003
0.2423 1.0984 67 0.4894 67.8341 21.0337
0.2817 1.1148 68 0.4882 67.5576 21.0603
0.3196 1.1311 69 0.4873 66.5438 20.5275
0.327 1.1475 70 0.4863 66.3594 20.5808
0.2701 1.1639 71 0.4863 67.5576 20.8072
0.2099 1.1803 72 0.4872 67.4654 20.9405
0.2558 1.1967 73 0.4882 67.7419 20.9405
0.2471 1.2131 74 0.4881 67.7419 20.9271
0.2207 1.2295 75 0.4878 67.6498 20.9138
0.2822 1.2459 76 0.4870 68.5714 20.9804
0.2616 1.2623 77 0.4864 68.7558 20.9271
0.278 1.2787 78 0.4855 69.1244 21.0470
0.2461 1.2951 79 0.4847 72.0737 21.9928
0.2284 1.3115 80 0.4844 71.6129 21.7264
0.2435 1.3279 81 0.4837 71.5207 21.6598
0.2686 1.3443 82 0.4833 67.3733 19.9547
0.2648 1.3607 83 0.4824 68.0184 20.1279
0.2469 1.3770 84 0.4819 68.0184 20.2611
0.2287 1.3934 85 0.4810 68.2949 20.0480
0.2344 1.4098 86 0.4809 67.7419 19.8748
0.2436 1.4262 87 0.4802 68.1106 19.8481
0.2397 1.4426 88 0.4801 67.0968 19.7283
0.2616 1.4590 89 0.4796 67.4654 19.8348
0.23 1.4754 90 0.4794 66.8203 19.8348
0.2726 1.4918 91 0.4781 67.0046 19.9281
0.2342 1.5082 92 0.4775 67.2811 19.9947
0.1631 1.5246 93 0.4766 67.4654 19.9147
0.2532 1.5410 94 0.4767 66.6359 19.8881
0.2411 1.5574 95 0.4757 66.3594 19.7016
0.2854 1.5738 96 0.4752 66.1751 19.7283
0.2687 1.5902 97 0.4742 66.5438 19.7549
0.2404 1.6066 98 0.4738 66.1751 19.7815
0.1993 1.6230 99 0.4733 66.3594 19.7682
0.213 1.6393 100 0.4729 66.4516 19.8082
0.1995 1.6557 101 0.4728 66.4516 19.7416
0.2291 1.6721 102 0.4730 66.8203 19.7949
0.2392 1.6885 103 0.4727 66.7281 19.8082
0.2617 1.7049 104 0.4724 66.7281 19.8748
0.3 1.7213 105 0.4725 66.6359 19.9014
0.2185 1.7377 106 0.4732 66.7281 20.1279
0.2271 1.7541 107 0.4735 66.6359 19.9147
0.1992 1.7705 108 0.4733 66.6359 19.9414
0.2123 1.7869 109 0.4738 66.6359 19.8615
0.2326 1.8033 110 0.4743 66.9124 19.8215
0.269 1.8197 111 0.4739 66.5438 19.7949
0.2053 1.8361 112 0.4739 66.7281 19.7815
0.2034 1.8525 113 0.4739 66.4516 19.7416
0.2033 1.8689 114 0.4739 66.4516 19.7682
0.2734 1.8852 115 0.4737 66.3594 19.7149
0.24 1.9016 116 0.4739 66.3594 19.7149
0.2612 1.9180 117 0.4737 66.6359 19.8082
0.2659 1.9344 118 0.4737 66.6359 19.8481
0.224 1.9508 119 0.4736 66.7281 19.8881
0.2204 1.9672 120 0.4736 66.8203 19.8215
0.2356 1.9836 121 0.4737 66.6359 19.7682
0.2193 2.0 122 0.4736 66.5438 19.8215

Framework versions

  • Transformers 4.57.3
  • Pytorch 2.9.1+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
1
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nkkbr/whisper-large-v3-zatoichi-ja-JDG_ver_20260212_lr_1.4e-5

Finetuned
(816)
this model