Whisper Large v3 - Japanese Zatoichi ASR

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4721
  • Wer: 66.6359
  • Cer: 20.0746

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.0467 0.0164 1 1.2545 98.1567 40.8419
1.1366 0.0328 2 0.9974 100.4608 39.9760
0.9623 0.0492 3 0.9468 92.9954 36.7524
0.8227 0.0656 4 0.8737 90.4147 33.9417
0.7999 0.0820 5 0.8299 86.7281 31.3041
0.7585 0.0984 6 0.7691 94.7465 38.0445
0.7779 0.1148 7 0.7384 91.7972 32.0767
0.787 0.1311 8 0.7099 101.1060 36.4859
0.5704 0.1475 9 0.6876 101.2903 38.5107
0.5429 0.1639 10 0.6649 96.1290 36.1396
0.6962 0.1803 11 0.6433 93.2719 31.6105
0.6661 0.1967 12 0.6266 94.1014 31.6238
0.6868 0.2131 13 0.6249 84.7005 28.0805
0.6043 0.2295 14 0.6109 79.3548 25.8292
0.5467 0.2459 15 0.6145 84.5161 28.6799
0.5758 0.2623 16 0.6086 79.0783 27.3212
0.4723 0.2787 17 0.5990 82.2120 34.0749
0.476 0.2951 18 0.5992 86.5438 41.0151
0.5545 0.3115 19 0.5883 85.4378 32.9692
0.4452 0.3279 20 0.5830 75.9447 26.7084
0.5191 0.3443 21 0.5775 76.0369 25.3763
0.5186 0.3607 22 0.5731 74.3779 25.7759
0.5398 0.3770 23 0.5813 74.0092 23.9776
0.5192 0.3934 24 0.5793 73.1797 24.3639
0.5424 0.4098 25 0.5664 72.9032 23.8977
0.4434 0.4262 26 0.5615 75.2995 23.6579
0.4001 0.4426 27 0.5614 75.4839 23.7778
0.4888 0.4590 28 0.5571 71.4286 23.1917
0.4519 0.4754 29 0.5642 70.8756 23.5247
0.4317 0.4918 30 0.5631 71.4286 23.6579
0.5509 0.5082 31 0.5596 71.8894 23.5913
0.5109 0.5246 32 0.5555 72.9954 23.4048
0.4176 0.5410 33 0.5491 72.8111 23.2716
0.5682 0.5574 34 0.5445 76.6820 23.6846
0.4805 0.5738 35 0.5391 75.7604 23.5913
0.4523 0.5902 36 0.5344 72.4424 22.8320
0.4466 0.6066 37 0.5338 70.9677 23.0052
0.5174 0.6230 38 0.5314 71.2442 22.8453
0.463 0.6393 39 0.5275 71.0599 22.4724
0.5266 0.6557 40 0.5233 68.9401 22.1527
0.4467 0.6721 41 0.5153 69.3088 22.0994
0.4784 0.6885 42 0.5122 70.7834 22.1393
0.4456 0.7049 43 0.5103 71.3364 21.9262
0.4364 0.7213 44 0.5081 71.5207 22.0194
0.3935 0.7377 45 0.5084 70.6912 22.1260
0.4374 0.7541 46 0.5074 70.3226 22.1127
0.4791 0.7705 47 0.5079 69.4931 22.5523
0.4756 0.7869 48 0.5087 69.4009 22.5523
0.3973 0.8033 49 0.5082 68.3871 22.0461
0.5047 0.8197 50 0.5066 69.4009 23.4714
0.4794 0.8361 51 0.5038 69.4009 23.2050
0.3947 0.8525 52 0.5013 68.2949 21.8862
0.5726 0.8689 53 0.5004 69.2166 22.3392
0.3449 0.8852 54 0.5001 69.4931 22.2326
0.4603 0.9016 55 0.5002 70.0461 22.5390
0.4423 0.9180 56 0.5014 74.1014 23.9510
0.3694 0.9344 57 0.5012 75.2074 24.6969
0.4133 0.9508 58 0.4972 74.1935 24.5105
0.4305 0.9672 59 0.4930 75.1152 24.9367
0.4464 0.9836 60 0.4914 71.4286 24.2973
0.3402 1.0 61 0.4947 75.5760 23.8444
0.3283 1.0164 62 0.4945 77.2350 24.8701
0.2436 1.0328 63 0.4921 72.9032 24.4705
0.1821 1.0492 64 0.4932 67.8341 22.1393
0.2508 1.0656 65 0.4963 67.3733 21.4333
0.2097 1.0820 66 0.4989 66.9124 21.1936
0.2275 1.0984 67 0.5002 67.0968 21.5799
0.2842 1.1148 68 0.4968 65.4378 20.6341
0.2982 1.1311 69 0.4932 66.5438 20.4609
0.3103 1.1475 70 0.4900 70.5991 21.8063
0.2519 1.1639 71 0.4897 67.7419 21.0470
0.1926 1.1803 72 0.4921 67.1889 20.9538
0.2351 1.1967 73 0.4948 67.5576 21.0071
0.2368 1.2131 74 0.4945 67.0968 20.9405
0.2021 1.2295 75 0.4935 71.9816 22.6189
0.2649 1.2459 76 0.4914 68.2028 21.1403
0.2413 1.2623 77 0.4887 71.6129 22.3258
0.2593 1.2787 78 0.4866 72.2581 22.2059
0.2202 1.2951 79 0.4857 69.1244 20.6341
0.2029 1.3115 80 0.4857 69.7696 20.6208
0.227 1.3279 81 0.4853 69.7696 20.5541
0.2448 1.3443 82 0.4845 69.4009 20.5142
0.2366 1.3607 83 0.4832 68.8479 20.4209
0.2253 1.3770 84 0.4822 69.5853 20.7140
0.2132 1.3934 85 0.4809 68.9401 20.2611
0.2218 1.4098 86 0.4795 73.4562 22.0727
0.2257 1.4262 87 0.4785 73.4562 21.9928
0.2143 1.4426 88 0.4777 67.5576 20.1012
0.2412 1.4590 89 0.4775 68.2949 20.4343
0.2081 1.4754 90 0.4762 67.2811 20.1146
0.2528 1.4918 91 0.4748 67.2811 20.0879
0.2184 1.5082 92 0.4740 67.2811 20.1012
0.1486 1.5246 93 0.4727 67.3733 20.2744
0.2317 1.5410 94 0.4724 66.5438 19.9814
0.2182 1.5574 95 0.4712 66.6359 20.1279
0.2537 1.5738 96 0.4705 66.8203 19.9414
0.2453 1.5902 97 0.4703 66.6359 20.0480
0.2181 1.6066 98 0.4697 67.1889 20.0613
0.1844 1.6230 99 0.4697 67.1889 19.9680
0.192 1.6393 100 0.4698 67.0968 19.9281
0.1787 1.6557 101 0.4697 66.9124 19.8748
0.2061 1.6721 102 0.4700 66.6359 19.9147
0.2218 1.6885 103 0.4703 66.6359 19.9814
0.2354 1.7049 104 0.4705 67.1889 20.1146
0.2686 1.7213 105 0.4706 66.9124 20.0480
0.1979 1.7377 106 0.4710 66.8203 20.0480
0.2075 1.7541 107 0.4712 66.9124 20.1146
0.1811 1.7705 108 0.4716 66.6359 20.1012
0.1903 1.7869 109 0.4722 66.4516 20.0480
0.2148 1.8033 110 0.4723 66.3594 20.0213
0.2501 1.8197 111 0.4722 66.4516 20.0480
0.185 1.8361 112 0.4723 66.2673 20.1012
0.1836 1.8525 113 0.4722 66.3594 20.0346
0.187 1.8689 114 0.4722 66.8203 20.2211
0.2457 1.8852 115 0.4721 66.4516 20.0613
0.2309 1.9016 116 0.4720 66.5438 20.1678
0.2469 1.9180 117 0.4717 66.2673 20.0746
0.2438 1.9344 118 0.4718 66.5438 20.0746
0.2073 1.9508 119 0.4720 66.8203 20.0746
0.1977 1.9672 120 0.4718 66.6359 20.0746
0.2149 1.9836 121 0.4717 66.6359 20.1279
0.2062 2.0 122 0.4721 66.6359 20.0746

Framework versions

  • Transformers 4.57.3
  • Pytorch 2.9.1+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
1
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nkkbr/whisper-large-v3-zatoichi-ja-JDG_ver_20260212_lr_2.0e-5

Finetuned
(816)
this model