Whisper Large v3 - Japanese Zatoichi ASR

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4738
  • Wer: 66.4516
  • Cer: 20.0613

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2.2e-05
  • train_batch_size: 32
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Wer Cer
1.0467 0.0164 1 1.2545 98.1567 40.8419
1.1366 0.0328 2 0.9969 100.4608 40.0559
0.9618 0.0492 3 0.9433 92.3502 36.4593
0.8201 0.0656 4 0.8657 89.1244 33.1024
0.7934 0.0820 5 0.8056 89.1244 30.9311
0.7371 0.0984 6 0.7576 92.7189 33.3156
0.7617 0.1148 7 0.7260 90.9677 34.2614
0.7717 0.1311 8 0.7016 109.8618 41.5212
0.5622 0.1475 9 0.6797 98.1567 37.2452
0.5371 0.1639 10 0.6584 98.1567 32.5430
0.6886 0.1803 11 0.6384 96.4977 32.8227
0.6591 0.1967 12 0.6277 84.8848 28.1870
0.6858 0.2131 13 0.6189 80.3687 26.1489
0.5996 0.2295 14 0.6120 77.7880 25.3896
0.545 0.2459 15 0.6140 89.0323 31.3974
0.5714 0.2623 16 0.6073 81.8433 31.2375
0.4672 0.2787 17 0.6029 88.9401 45.7307
0.4809 0.2951 18 0.5958 85.7143 35.6734
0.5443 0.3115 19 0.5939 88.9401 29.7056
0.445 0.3279 20 0.5891 76.5899 24.4305
0.5222 0.3443 21 0.5787 81.7512 27.1347
0.5174 0.3607 22 0.5783 77.1429 28.8797
0.5426 0.3770 23 0.5919 74.3779 24.7103
0.533 0.3934 24 0.5825 72.8111 24.3106
0.5415 0.4098 25 0.5680 72.0737 23.4448
0.4445 0.4262 26 0.5652 82.1198 27.7741
0.4043 0.4426 27 0.5666 79.8157 27.1746
0.4958 0.4590 28 0.5567 70.9677 22.9119
0.4539 0.4754 29 0.5622 71.3364 23.6180
0.4294 0.4918 30 0.5668 70.7834 23.4581
0.5555 0.5082 31 0.5568 71.3364 23.1118
0.5043 0.5246 32 0.5463 72.5346 23.1118
0.4138 0.5410 33 0.5447 72.1659 23.1118
0.5678 0.5574 34 0.5419 75.1152 23.4315
0.4812 0.5738 35 0.5360 73.5484 23.2050
0.4521 0.5902 36 0.5327 72.1659 22.8986
0.4489 0.6066 37 0.5300 71.4286 22.8720
0.5172 0.6230 38 0.5264 69.9539 22.3924
0.458 0.6393 39 0.5194 73.1797 25.1632
0.525 0.6557 40 0.5153 69.3088 21.9262
0.4417 0.6721 41 0.5105 68.9401 22.0328
0.471 0.6885 42 0.5102 69.4009 22.2592
0.4417 0.7049 43 0.5097 69.4009 21.8596
0.4368 0.7213 44 0.5084 70.0461 21.9395
0.3992 0.7377 45 0.5051 69.5853 22.1527
0.432 0.7541 46 0.5032 69.8618 22.2059
0.4762 0.7705 47 0.5038 68.5714 22.2459
0.4783 0.7869 48 0.5057 68.0184 22.0328
0.3978 0.8033 49 0.5060 68.6636 22.0061
0.5011 0.8197 50 0.5050 68.4793 22.0994
0.4803 0.8361 51 0.5031 68.6636 22.2592
0.399 0.8525 52 0.5008 69.1244 21.9928
0.5721 0.8689 53 0.5000 69.5853 22.3791
0.3436 0.8852 54 0.4980 68.9401 22.0861
0.4632 0.9016 55 0.4961 69.1244 22.1260
0.443 0.9180 56 0.4964 72.6267 23.6046
0.3709 0.9344 57 0.4975 73.8249 23.8178
0.4071 0.9508 58 0.4942 74.8387 24.7502
0.4331 0.9672 59 0.4905 73.9171 24.4971
0.4466 0.9836 60 0.4898 68.4793 21.6731
0.3378 1.0 61 0.4930 81.5668 26.1889
0.3231 1.0164 62 0.4928 78.2488 24.5904
0.2373 1.0328 63 0.4911 70.9677 23.0984
0.1737 1.0492 64 0.4970 70.3226 23.2849
0.2477 1.0656 65 0.5023 68.5714 21.8330
0.2072 1.0820 66 0.5022 67.4654 21.4466
0.2245 1.0984 67 0.5002 71.2442 22.6855
0.2653 1.1148 68 0.4984 65.9908 20.5808
0.2913 1.1311 69 0.4974 68.2949 21.1536
0.3033 1.1475 70 0.4950 69.6774 21.4733
0.2467 1.1639 71 0.4942 71.3364 22.7654
0.1891 1.1803 72 0.4967 66.4516 21.3001
0.225 1.1967 73 0.4992 67.4654 21.1936
0.2322 1.2131 74 0.4981 66.8203 20.8872
0.1976 1.2295 75 0.4960 71.8894 22.6056
0.2555 1.2459 76 0.4942 71.5207 22.5123
0.2383 1.2623 77 0.4917 68.4793 20.7273
0.2498 1.2787 78 0.4891 68.3871 20.5275
0.2148 1.2951 79 0.4879 69.3088 20.7007
0.1962 1.3115 80 0.4878 68.5714 20.4076
0.2198 1.3279 81 0.4868 68.0184 20.3277
0.2388 1.3443 82 0.4868 68.2949 20.3543
0.2295 1.3607 83 0.4852 68.7558 20.5142
0.2234 1.3770 84 0.4834 69.4009 20.6740
0.2089 1.3934 85 0.4819 69.2166 20.6208
0.2101 1.4098 86 0.4806 72.9032 22.1660
0.2202 1.4262 87 0.4794 73.0876 22.2859
0.213 1.4426 88 0.4787 72.3502 22.1793
0.2289 1.4590 89 0.4781 67.6498 20.5275
0.2059 1.4754 90 0.4778 67.7419 20.4209
0.2496 1.4918 91 0.4764 67.2811 20.2611
0.2136 1.5082 92 0.4752 67.7419 20.4343
0.1397 1.5246 93 0.4749 67.2811 20.3543
0.2214 1.5410 94 0.4744 67.5576 20.3277
0.2123 1.5574 95 0.4733 67.4654 20.4609
0.2489 1.5738 96 0.4730 67.8341 20.3144
0.2405 1.5902 97 0.4729 66.9124 20.0879
0.2122 1.6066 98 0.4724 66.7281 20.2344
0.1745 1.6230 99 0.4720 67.0046 20.1012
0.1881 1.6393 100 0.4719 66.7281 20.0613
0.1737 1.6557 101 0.4722 66.9124 20.1945
0.1997 1.6721 102 0.4725 67.8341 20.3277
0.2096 1.6885 103 0.4726 66.8203 20.3410
0.2266 1.7049 104 0.4729 67.0046 20.2478
0.2651 1.7213 105 0.4726 66.1751 20.1279
0.1875 1.7377 106 0.4734 66.3594 20.2877
0.2011 1.7541 107 0.4739 66.5438 20.2478
0.1769 1.7705 108 0.4744 66.3594 20.1412
0.1861 1.7869 109 0.4743 66.3594 20.0346
0.1993 1.8033 110 0.4751 66.3594 20.0346
0.2387 1.8197 111 0.4747 67.0046 20.2344
0.1803 1.8361 112 0.4750 66.2673 20.0213
0.1784 1.8525 113 0.4746 66.8203 20.1146
0.1794 1.8689 114 0.4744 66.1751 20.0080
0.2426 1.8852 115 0.4743 66.7281 20.0746
0.2283 1.9016 116 0.4743 66.7281 20.1146
0.2412 1.9180 117 0.4741 66.6359 20.0879
0.2372 1.9344 118 0.4738 66.6359 20.0746
0.2024 1.9508 119 0.4736 66.8203 20.1279
0.195 1.9672 120 0.4739 66.9124 20.1545
0.2083 1.9836 121 0.4738 66.9124 20.1945
0.1995 2.0 122 0.4738 66.4516 20.0613

Framework versions

  • Transformers 4.57.3
  • Pytorch 2.9.1+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
1
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nkkbr/whisper-large-v3-zatoichi-ja-JDG_ver_20260212_lr_2.2e-5

Finetuned
(816)
this model