Whisper Large v3 - Japanese Zatoichi ASR

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3878
  • Wer: 73.0876
  • Cer: 19.3819

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2.4e-05
  • train_batch_size: 32
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.2734 0.0065 1 1.2545 98.1567 40.8419
1.357 0.0130 2 0.9966 95.7604 38.4574
1.1204 0.0195 3 0.9725 93.8249 37.2186
0.9379 0.0260 4 0.9338 90.1382 35.6467
1.0264 0.0325 5 0.9038 86.5438 33.9283
0.8522 0.0390 6 0.8448 89.4009 32.0767
0.7404 0.0455 7 0.8027 92.9032 32.3565
0.8627 0.0519 8 0.7624 104.1475 40.1892
0.7296 0.0584 9 0.7394 93.5484 31.1842
0.7422 0.0649 10 0.7100 102.2120 39.1235
0.724 0.0714 11 0.6762 113.4562 41.0151
0.8729 0.0779 12 0.6532 101.0138 37.4850
0.5875 0.0844 13 0.6302 82.4885 27.9073
0.5907 0.0909 14 0.6158 92.7189 29.5458
0.6703 0.0974 15 0.6062 88.7558 27.6675
0.7156 0.1039 16 0.5941 78.1567 24.2707
0.629 0.1104 17 0.5893 78.8940 23.8577
0.5721 0.1169 18 0.5869 78.5253 23.2716
0.524 0.1234 19 0.5774 85.8065 26.4686
0.6045 0.1299 20 0.5705 86.6359 26.9215
0.6053 0.1364 21 0.5731 84.0553 26.0823
0.6037 0.1429 22 0.5621 77.0507 23.4181
0.7245 0.1494 23 0.5556 82.2120 25.6294
0.5523 0.1558 24 0.5508 77.2350 23.4581
0.535 0.1623 25 0.5529 78.4332 25.1365
0.8656 0.1688 26 0.5512 78.0645 24.7369
0.5497 0.1753 27 0.5380 74.4700 23.2317
0.5567 0.1818 28 0.5414 76.5899 26.0290
0.5823 0.1883 29 0.5314 70.9677 25.3497
0.6001 0.1948 30 0.5320 69.4009 27.8140
0.6098 0.2013 31 0.5384 72.1659 30.0520
0.5954 0.2078 32 0.5260 76.1290 26.2422
0.4836 0.2143 33 0.5277 82.6728 24.1908
0.463 0.2208 34 0.5335 78.3410 23.0585
0.6094 0.2273 35 0.5336 72.5346 22.4724
0.5365 0.2338 36 0.5294 72.1659 22.2059
0.5999 0.2403 37 0.5182 78.7097 23.8977
0.5659 0.2468 38 0.5171 75.8525 22.5922
0.642 0.2532 39 0.5130 74.2857 22.2459
0.5999 0.2597 40 0.5138 78.9862 25.4030
0.6069 0.2662 41 0.5039 71.4286 22.0861
0.4999 0.2727 42 0.4980 66.6359 21.0071
0.4597 0.2792 43 0.5054 69.5853 21.3534
0.6185 0.2857 44 0.5036 75.2995 23.5247
0.4341 0.2922 45 0.4992 77.7880 23.6846
0.4685 0.2987 46 0.5002 76.9585 21.9928
0.5824 0.3052 47 0.5004 80.7373 23.1251
0.5148 0.3117 48 0.5003 80.2765 22.9253
0.5157 0.3182 49 0.4975 75.5760 22.3258
0.5239 0.3247 50 0.4949 72.3502 21.8596
0.4329 0.3312 51 0.4924 68.4793 20.8206
0.5947 0.3377 52 0.4938 70.5069 21.6598
0.5355 0.3442 53 0.4950 71.7972 22.1793
0.4978 0.3506 54 0.4878 79.1705 25.5495
0.4475 0.3571 55 0.4865 82.3963 25.1765
0.5051 0.3636 56 0.4920 89.3088 27.5476
0.4721 0.3701 57 0.4876 85.3456 26.0424
0.5862 0.3766 58 0.4792 78.8940 23.9643
0.4881 0.3831 59 0.4775 70.1382 20.9005
0.4749 0.3896 60 0.4817 68.3871 20.6341
0.5812 0.3961 61 0.4813 68.3871 21.0337
0.5496 0.4026 62 0.4777 70.4147 21.1536
0.6596 0.4091 63 0.4754 80.3687 23.2317
0.4739 0.4156 64 0.4755 83.2258 23.1784
0.509 0.4221 65 0.4718 82.1198 24.4305
0.4443 0.4286 66 0.4693 76.3134 23.1650
0.4901 0.4351 67 0.4709 72.6267 23.0185
0.4674 0.4416 68 0.4715 70.6912 21.4067
0.3878 0.4481 69 0.4748 74.7465 23.7778
0.4964 0.4545 70 0.4791 73.2719 21.7530
0.5233 0.4610 71 0.4804 77.3272 23.6313
0.6095 0.4675 72 0.4801 76.1290 23.3249
0.475 0.4740 73 0.4746 74.3779 22.5656
0.5331 0.4805 74 0.4669 70.2304 20.9538
0.4976 0.4870 75 0.4678 68.5714 20.3144
0.5064 0.4935 76 0.4710 73.2719 22.6322
0.4759 0.5 77 0.4717 82.0276 25.8292
0.4829 0.5065 78 0.4680 79.4470 25.2431
0.4745 0.5130 79 0.4636 79.5392 23.3382
0.4536 0.5195 80 0.4595 76.4977 23.8577
0.5633 0.5260 81 0.4585 78.4332 23.5514
0.4319 0.5325 82 0.4590 77.9724 23.2716
0.4888 0.5390 83 0.4575 76.0369 22.9386
0.5005 0.5455 84 0.4560 70.5069 22.0727
0.4933 0.5519 85 0.4572 70.0461 22.0194
0.4965 0.5584 86 0.4609 74.3779 24.1108
0.4328 0.5649 87 0.4628 75.7604 24.3639
0.6036 0.5714 88 0.4625 81.1982 26.5885
0.5975 0.5779 89 0.4588 81.9355 26.5486
0.3809 0.5844 90 0.4547 85.4378 25.1099
0.4017 0.5909 91 0.4534 93.0876 28.5334
0.4532 0.5974 92 0.4519 76.2212 22.7255
0.5179 0.6039 93 0.4509 71.7972 21.8463
0.4977 0.6104 94 0.4509 70.5991 21.7131
0.4844 0.6169 95 0.4510 65.7143 19.5817
0.5256 0.6234 96 0.4534 68.4793 20.2211
0.3784 0.6299 97 0.4585 69.4009 20.1279
0.5134 0.6364 98 0.4582 72.1659 21.0737
0.5261 0.6429 99 0.4523 70.5991 20.5675
0.491 0.6494 100 0.4473 72.8111 20.9005
0.479 0.6558 101 0.4441 76.6820 23.0718
0.3678 0.6623 102 0.4413 76.1290 22.8587
0.4049 0.6688 103 0.4375 74.1935 21.8596
0.4442 0.6753 104 0.4382 73.6406 21.5266
0.4573 0.6818 105 0.4379 72.9032 21.7397
0.5421 0.6883 106 0.4370 68.6636 20.3677
0.3765 0.6948 107 0.4385 68.3871 20.6208
0.3816 0.7013 108 0.4368 68.7558 19.9947
0.4938 0.7078 109 0.4332 70.3226 20.5009
0.5335 0.7143 110 0.4309 76.7742 24.0176
0.5403 0.7208 111 0.4305 79.6313 24.7103
0.3915 0.7273 112 0.4306 77.6037 23.8444
0.5789 0.7338 113 0.4283 72.6267 21.2735
0.4838 0.7403 114 0.4248 72.4424 21.2069
0.4032 0.7468 115 0.4255 72.9954 20.9271
0.4089 0.7532 116 0.4262 75.2995 21.6997
0.4211 0.7597 117 0.4236 68.9401 19.9414
0.5029 0.7662 118 0.4191 67.9263 19.6217
0.5466 0.7727 119 0.4164 67.3733 19.4618
0.4455 0.7792 120 0.4146 69.6774 20.5009
0.4386 0.7857 121 0.4143 66.5438 19.3419
0.5332 0.7922 122 0.4131 67.4654 18.9290
0.4302 0.7987 123 0.4109 72.9032 20.5541
0.494 0.8052 124 0.4125 73.6406 20.3410
0.4382 0.8117 125 0.4131 73.0876 20.3144
0.523 0.8182 126 0.4133 70.6912 20.7540
0.4775 0.8247 127 0.4133 72.5346 21.8330
0.4565 0.8312 128 0.4160 68.6636 20.5675
0.4969 0.8377 129 0.4183 71.9816 22.6589
0.5679 0.8442 130 0.4189 72.4424 23.0718
0.4157 0.8506 131 0.4197 69.8618 21.9395
0.4114 0.8571 132 0.4189 72.3502 22.3525
0.4229 0.8636 133 0.4191 72.2581 22.6322
0.4696 0.8701 134 0.4205 74.1014 22.7255
0.4032 0.8766 135 0.4228 74.3779 22.7121
0.4288 0.8831 136 0.4271 77.0507 22.2859
0.4184 0.8896 137 0.4282 73.9171 21.0870
0.4935 0.8961 138 0.4270 73.0876 22.1127
0.4234 0.9026 139 0.4253 70.6912 21.2468
0.4181 0.9091 140 0.4247 69.1244 21.0470
0.4259 0.9156 141 0.4254 67.5576 19.2487
0.3989 0.9221 142 0.4278 68.3871 19.5817
0.5154 0.9286 143 0.4306 73.5484 21.4999
0.3884 0.9351 144 0.4334 75.4839 22.8986
0.4956 0.9416 145 0.4339 75.2995 21.6731
0.3983 0.9481 146 0.4325 72.3502 20.5808
0.3428 0.9545 147 0.4317 72.5346 20.5275
0.4288 0.9610 148 0.4291 72.9954 20.4476
0.4419 0.9675 149 0.4261 71.4286 19.5684
0.3963 0.9740 150 0.4234 70.9677 19.3686
0.3982 0.9805 151 0.4201 70.4147 18.8757
0.4878 0.9870 152 0.4176 69.1244 18.4228
0.3999 0.9935 153 0.4139 67.5576 18.0365
0.44 1.0 154 0.4126 66.0829 17.9965
0.2352 1.0065 155 0.4120 65.1613 17.8900
0.2563 1.0130 156 0.4126 65.0691 18.2230
0.252 1.0195 157 0.4129 64.4240 18.1431
0.2941 1.0260 158 0.4145 66.3594 18.6759
0.2975 1.0325 159 0.4161 67.3733 18.5427
0.3141 1.0390 160 0.4181 66.5438 18.5161
0.2628 1.0455 161 0.4201 65.9908 18.4494
0.2799 1.0519 162 0.4233 67.7419 18.6226
0.2924 1.0584 163 0.4265 69.7696 19.0089
0.3154 1.0649 164 0.4287 71.0599 19.0622
0.2542 1.0714 165 0.4307 71.8894 19.0489
0.2766 1.0779 166 0.4281 71.4286 19.1022
0.2514 1.0844 167 0.4255 71.1521 19.1821
0.2641 1.0909 168 0.4232 71.6129 19.3020
0.2932 1.0974 169 0.4225 71.6129 18.9956
0.2569 1.1039 170 0.4192 73.6406 19.9547
0.2089 1.1104 171 0.4177 74.0092 20.1279
0.2565 1.1169 172 0.4160 72.9032 20.2344
0.2523 1.1234 173 0.4146 72.7189 20.1545
0.286 1.1299 174 0.4118 79.0783 22.1527
0.2086 1.1364 175 0.4088 68.1106 18.3962
0.2537 1.1429 176 0.4084 67.9263 18.1164
0.2241 1.1494 177 0.4088 68.2949 18.3562
0.21 1.1558 178 0.4105 68.1106 18.8757
0.2266 1.1623 179 0.4120 72.9032 20.7673
0.2159 1.1688 180 0.4131 67.8341 18.8358
0.2979 1.1753 181 0.4127 67.3733 18.8624
0.2714 1.1818 182 0.4140 70.8756 20.7273
0.1959 1.1883 183 0.4146 71.7972 20.7007
0.2589 1.1948 184 0.4140 66.3594 19.0089
0.2617 1.2013 185 0.4128 68.7558 18.8091
0.1976 1.2078 186 0.4135 69.6774 18.4894
0.2816 1.2143 187 0.4129 71.6129 18.7425
0.2766 1.2208 188 0.4112 71.2442 18.4628
0.2858 1.2273 189 0.4082 71.1521 18.3562
0.2625 1.2338 190 0.4027 67.1889 18.1031
0.2492 1.2403 191 0.4002 66.4516 18.1031
0.2665 1.2468 192 0.3992 66.9124 18.1697
0.2219 1.2532 193 0.3994 66.9124 18.4761
0.2664 1.2597 194 0.3983 67.4654 18.6626
0.2267 1.2662 195 0.3979 68.0184 18.7558
0.2375 1.2727 196 0.3986 69.8618 18.9024
0.2478 1.2792 197 0.3999 70.8756 18.9556
0.2104 1.2857 198 0.3999 70.3226 18.7425
0.2476 1.2922 199 0.3999 69.2166 18.5427
0.2228 1.2987 200 0.3986 69.1244 18.6093
0.2238 1.3052 201 0.3986 68.2028 18.8224
0.2712 1.3117 202 0.3984 66.3594 18.3962
0.289 1.3182 203 0.3990 66.1751 18.4494
0.2181 1.3247 204 0.3997 65.6221 18.3296
0.2462 1.3312 205 0.4002 65.9908 18.2896
0.278 1.3377 206 0.4007 66.5438 18.2630
0.2184 1.3442 207 0.4023 68.6636 18.4628
0.2391 1.3506 208 0.4025 69.2166 18.6626
0.2612 1.3571 209 0.4031 69.0323 18.5960
0.251 1.3636 210 0.4029 68.8479 18.5960
0.2653 1.3701 211 0.4027 67.6498 18.4095
0.2094 1.3766 212 0.4017 66.8203 18.3828
0.2223 1.3831 213 0.4018 65.8986 18.2496
0.2145 1.3896 214 0.4022 67.0046 18.2230
0.2016 1.3961 215 0.4029 66.2673 18.0765
0.2321 1.4026 216 0.4033 66.6359 18.1031
0.2485 1.4091 217 0.4035 67.2811 18.2630
0.2229 1.4156 218 0.4045 68.7558 18.2630
0.2313 1.4221 219 0.4057 71.0599 18.5960
0.2114 1.4286 220 0.4083 71.7051 18.6359
0.2315 1.4351 221 0.4095 72.6267 18.8224
0.207 1.4416 222 0.4104 74.1935 18.8757
0.2522 1.4481 223 0.4096 73.9171 18.7958
0.2369 1.4545 224 0.4079 73.2719 18.9423
0.232 1.4610 225 0.4066 71.7051 18.6626
0.178 1.4675 226 0.4053 71.7972 18.7558
0.2096 1.4740 227 0.4027 68.8479 18.4095
0.2437 1.4805 228 0.4006 66.8203 18.2630
0.235 1.4870 229 0.3994 66.0829 17.9699
0.325 1.4935 230 0.3985 65.7143 17.9033
0.2288 1.5 231 0.3983 65.6221 18.0099
0.239 1.5065 232 0.3982 65.8986 18.0365
0.21 1.5130 233 0.3985 66.7281 18.2097
0.2485 1.5195 234 0.3992 68.4793 18.4228
0.2693 1.5260 235 0.3996 69.4931 18.5027
0.2264 1.5325 236 0.4004 70.6912 18.8757
0.2395 1.5390 237 0.4006 70.7834 18.9157
0.3003 1.5455 238 0.4002 71.4286 18.8624
0.1887 1.5519 239 0.3995 70.2304 18.5294
0.2644 1.5584 240 0.3996 71.8894 18.8491
0.2579 1.5649 241 0.3993 71.1521 18.7825
0.2582 1.5714 242 0.3989 71.3364 18.8091
0.1914 1.5779 243 0.3983 70.2304 18.6626
0.2378 1.5844 244 0.3974 69.6774 18.6493
0.1753 1.5909 245 0.3960 69.5853 18.5960
0.1921 1.5974 246 0.3959 70.0461 18.7292
0.2542 1.6039 247 0.3949 69.4009 18.7025
0.2528 1.6104 248 0.3946 69.5853 18.5560
0.198 1.6169 249 0.3934 69.5853 18.5827
0.1862 1.6234 250 0.3929 68.6636 18.4228
0.2556 1.6299 251 0.3918 68.4793 18.3562
0.2192 1.6364 252 0.3910 68.9401 18.4494
0.2243 1.6429 253 0.3904 68.4793 18.4228
0.2034 1.6494 254 0.3901 68.1106 18.2630
0.2827 1.6558 255 0.3903 67.9263 18.3162
0.1673 1.6623 256 0.3909 68.4793 18.0898
0.2625 1.6688 257 0.3906 68.2949 18.1431
0.2007 1.6753 258 0.3912 68.6636 18.1964
0.2523 1.6818 259 0.3914 68.8479 18.3029
0.2115 1.6883 260 0.3920 69.4009 18.3828
0.2244 1.6948 261 0.3922 70.6912 18.5427
0.254 1.7013 262 0.3918 70.7834 18.5027
0.3353 1.7078 263 0.3920 70.2304 18.4228
0.2417 1.7143 264 0.3917 69.7696 18.3695
0.2573 1.7208 265 0.3907 69.3088 18.2630
0.2464 1.7273 266 0.3904 69.2166 18.2496
0.2189 1.7338 267 0.3902 69.0323 18.2763
0.2141 1.7403 268 0.3898 68.6636 18.1564
0.1926 1.7468 269 0.3892 72.2581 19.5018
0.203 1.7532 270 0.3885 71.3364 19.4485
0.1783 1.7597 271 0.3884 71.7051 19.4752
0.2091 1.7662 272 0.3882 72.1659 19.3553
0.2924 1.7727 273 0.3886 68.6636 17.9433
0.2694 1.7792 274 0.3886 69.1244 17.8234
0.2207 1.7857 275 0.3885 69.1244 17.8100
0.1798 1.7922 276 0.3886 69.3088 17.8766
0.207 1.7987 277 0.3885 69.4009 17.8234
0.2635 1.8052 278 0.3891 69.6774 17.9433
0.2439 1.8117 279 0.3890 69.9539 17.8766
0.226 1.8182 280 0.3892 70.4147 18.0232
0.2634 1.8247 281 0.3893 70.0461 17.9965
0.2195 1.8312 282 0.3900 70.1382 17.9832
0.2109 1.8377 283 0.3895 70.1382 17.8766
0.23 1.8442 284 0.3895 69.7696 17.8766
0.2768 1.8506 285 0.3893 73.0876 19.3286
0.1995 1.8571 286 0.3889 73.3641 19.3686
0.2582 1.8636 287 0.3891 69.9539 17.9299
0.2292 1.8701 288 0.3888 69.8618 18.0365
0.2536 1.8766 289 0.3883 69.9539 17.9832
0.266 1.8831 290 0.3885 72.9954 19.2087
0.2385 1.8896 291 0.3882 72.9032 19.2354
0.2365 1.8961 292 0.3880 72.7189 19.2887
0.1852 1.9026 293 0.3880 73.0876 19.3553
0.2181 1.9091 294 0.3877 73.1797 19.3819
0.2273 1.9156 295 0.3879 72.9954 19.3153
0.2787 1.9221 296 0.3879 72.7189 19.2620
0.277 1.9286 297 0.3880 72.6267 19.2887
0.2341 1.9351 298 0.3878 72.9032 19.3553
0.2629 1.9416 299 0.3880 72.8111 19.3286
0.2252 1.9481 300 0.3879 72.7189 19.3286
0.2619 1.9545 301 0.3878 72.9954 19.3553
0.2488 1.9610 302 0.3877 72.6267 19.3553
0.267 1.9675 303 0.3874 73.1797 19.4752
0.2343 1.9740 304 0.3877 72.8111 19.3286
0.269 1.9805 305 0.3876 73.1797 19.3952
0.1681 1.9870 306 0.3878 72.9032 19.3286
0.2602 1.9935 307 0.3876 72.9954 19.3553
0.2112 2.0 308 0.3878 73.0876 19.3819

Framework versions

  • Transformers 4.57.3
  • Pytorch 2.9.1+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
6
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nkkbr/whisper-large-v3-zatoichi-ja-JDG_ver_20260220_lr_2.4e-5

Finetuned
(815)
this model