Whisper Large v3 - Japanese Zatoichi ASR

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4699
  • Wer: 66.4516
  • Cer: 20.0080

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1.8e-05
  • train_batch_size: 32
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.0467 0.0164 1 1.2545 98.1567 40.8419
1.1366 0.0328 2 0.9975 98.9862 38.6040
0.9623 0.0492 3 0.9529 93.7327 36.9522
0.8283 0.0656 4 0.8875 89.9539 34.8475
0.8133 0.0820 5 0.8385 89.1244 32.8627
0.7654 0.0984 6 0.7780 90.3226 31.5439
0.7871 0.1148 7 0.7467 90.4147 32.1300
0.7974 0.1311 8 0.7176 91.8894 35.9931
0.5786 0.1475 9 0.7020 102.7650 38.8970
0.5542 0.1639 10 0.6792 104.0553 40.2424
0.7073 0.1803 11 0.6534 96.0369 31.8636
0.6749 0.1967 12 0.6350 102.8571 34.3546
0.698 0.2131 13 0.6320 90.2304 30.9311
0.6138 0.2295 14 0.6138 76.5899 25.0566
0.5518 0.2459 15 0.6133 77.2350 25.4163
0.5693 0.2623 16 0.6070 77.5115 25.1365
0.4768 0.2787 17 0.5961 83.2258 28.7731
0.4734 0.2951 18 0.5965 80.1843 39.4831
0.5552 0.3115 19 0.5867 82.9493 32.4231
0.4442 0.3279 20 0.5810 82.4885 28.9730
0.5199 0.3443 21 0.5754 75.0230 23.9510
0.5181 0.3607 22 0.5692 74.5622 25.4962
0.533 0.3770 23 0.5814 73.6406 27.0814
0.5214 0.3934 24 0.5788 74.6544 27.5609
0.5405 0.4098 25 0.5651 73.3641 24.1508
0.445 0.4262 26 0.5583 75.1152 23.9243
0.3923 0.4426 27 0.5602 75.0230 23.5380
0.4868 0.4590 28 0.5534 77.2350 24.5904
0.4493 0.4754 29 0.5584 71.1521 23.1784
0.4254 0.4918 30 0.5595 71.7051 23.4581
0.5473 0.5082 31 0.5574 71.4286 23.3649
0.5057 0.5246 32 0.5543 72.4424 23.0984
0.415 0.5410 33 0.5503 73.9171 23.3249
0.5682 0.5574 34 0.5454 75.2995 23.4981
0.4804 0.5738 35 0.5396 76.2212 23.5247
0.4547 0.5902 36 0.5336 71.9816 22.8853
0.4501 0.6066 37 0.5299 78.6175 25.5228
0.5138 0.6230 38 0.5276 76.9585 25.0433
0.4667 0.6393 39 0.5244 71.3364 22.6056
0.523 0.6557 40 0.5210 73.7327 25.0300
0.4442 0.6721 41 0.5141 73.5484 24.9900
0.4771 0.6885 42 0.5111 71.6129 22.5390
0.4482 0.7049 43 0.5083 72.5346 22.5390
0.4369 0.7213 44 0.5044 71.4286 22.4191
0.3939 0.7377 45 0.5027 71.0599 22.3525
0.4351 0.7541 46 0.5024 70.5069 22.0061
0.4768 0.7705 47 0.5037 68.6636 22.0994
0.4756 0.7869 48 0.5045 69.2166 22.5390
0.3988 0.8033 49 0.5049 69.0323 22.2992
0.5043 0.8197 50 0.5037 69.1244 22.0061
0.4753 0.8361 51 0.5019 68.0184 21.8729
0.393 0.8525 52 0.4997 68.2028 21.6331
0.5722 0.8689 53 0.4982 69.0323 21.8862
0.3456 0.8852 54 0.4955 68.3871 21.9129
0.4647 0.9016 55 0.4938 72.7189 23.7512
0.4431 0.9180 56 0.4946 73.7327 23.4847
0.3653 0.9344 57 0.4964 72.7189 23.3249
0.4116 0.9508 58 0.4937 71.7972 23.2183
0.4366 0.9672 59 0.4886 72.8111 23.2583
0.4465 0.9836 60 0.4842 72.3502 24.3906
0.3432 1.0 61 0.4849 73.2719 24.3906
0.3308 1.0164 62 0.4853 74.3779 23.4581
0.2445 1.0328 63 0.4850 68.6636 21.6331
0.1849 1.0492 64 0.4877 68.8479 22.2326
0.2543 1.0656 65 0.4912 68.4793 22.0328
0.2168 1.0820 66 0.4925 67.9263 21.4866
0.2329 1.0984 67 0.4915 66.5438 21.1802
0.2677 1.1148 68 0.4898 70.5991 22.3392
0.3012 1.1311 69 0.4879 65.4378 20.5009
0.3084 1.1475 70 0.4864 66.6359 20.3810
0.2551 1.1639 71 0.4862 66.5438 20.6874
0.1999 1.1803 72 0.4880 66.9124 20.5941
0.2429 1.1967 73 0.4898 66.9124 20.9271
0.2409 1.2131 74 0.4893 67.0046 20.6740
0.2062 1.2295 75 0.4879 71.2442 22.2859
0.2622 1.2459 76 0.4873 70.7834 22.0594
0.2466 1.2623 77 0.4851 71.2442 21.8862
0.2619 1.2787 78 0.4835 71.5207 22.0328
0.2295 1.2951 79 0.4821 71.7051 21.9928
0.2092 1.3115 80 0.4818 72.5346 22.1260
0.23 1.3279 81 0.4814 69.2166 20.8072
0.2476 1.3443 82 0.4809 68.8479 20.7007
0.2425 1.3607 83 0.4799 68.2028 20.5675
0.2293 1.3770 84 0.4792 68.8479 20.7140
0.2174 1.3934 85 0.4779 72.9032 21.9528
0.2208 1.4098 86 0.4781 72.9954 22.0727
0.232 1.4262 87 0.4769 68.5714 20.3943
0.2224 1.4426 88 0.4759 68.2028 20.3410
0.2432 1.4590 89 0.4756 68.3871 20.3277
0.2128 1.4754 90 0.4752 68.1106 20.4476
0.2592 1.4918 91 0.4744 67.2811 20.1812
0.2232 1.5082 92 0.4735 67.1889 20.1279
0.1532 1.5246 93 0.4727 66.3594 19.9680
0.2344 1.5410 94 0.4715 66.6359 20.0746
0.2243 1.5574 95 0.4708 67.0968 20.3144
0.2635 1.5738 96 0.4699 66.2673 20.2211
0.2494 1.5902 97 0.4695 66.0829 20.1012
0.2258 1.6066 98 0.4694 65.9908 20.2211
0.1882 1.6230 99 0.4688 66.5438 20.1945
0.2084 1.6393 100 0.4691 66.8203 20.1279
0.1919 1.6557 101 0.4688 67.2811 20.2478
0.2097 1.6721 102 0.4689 66.7281 20.1678
0.2217 1.6885 103 0.4690 66.4516 20.0746
0.2423 1.7049 104 0.4693 66.3594 20.1279
0.2825 1.7213 105 0.4692 66.1751 20.0480
0.2083 1.7377 106 0.4694 66.0829 20.1279
0.2145 1.7541 107 0.4696 65.8065 20.0879
0.1843 1.7705 108 0.4700 66.3594 20.1012
0.2057 1.7869 109 0.4703 65.9908 20.0613
0.2141 1.8033 110 0.4704 66.3594 20.0346
0.2492 1.8197 111 0.4706 66.4516 20.0213
0.1897 1.8361 112 0.4707 66.5438 20.0613
0.1877 1.8525 113 0.4705 67.2811 20.2344
0.1913 1.8689 114 0.4702 67.0046 20.2211
0.2543 1.8852 115 0.4702 66.4516 20.0613
0.2303 1.9016 116 0.4701 66.6359 20.0080
0.2524 1.9180 117 0.4699 66.2673 19.9414
0.2503 1.9344 118 0.4699 66.4516 20.1412
0.2075 1.9508 119 0.4698 66.3594 20.0613
0.2046 1.9672 120 0.4698 66.4516 20.0613
0.2199 1.9836 121 0.4698 66.1751 20.0213
0.2092 2.0 122 0.4699 66.4516 20.0080

Framework versions

  • Transformers 4.57.3
  • Pytorch 2.9.1+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
1
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nkkbr/whisper-large-v3-zatoichi-ja-JDG_ver_20260212_lr_1.8e-5

Finetuned
(817)
this model