Configuration Parsing Warning:In adapter_config.json: "peft.task_type" must be a string

karato-whisper-latin

This model is a fine-tuned version of openai/whisper-large-v2 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7369
  • Wer: 0.6002
  • Cer: 0.2809

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 0.1
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Wer Cer
2.1077 0.1433 100 2.0875 1.1545 0.6442
1.3038 0.2865 200 1.2774 0.9410 0.4603
1.1622 0.4298 300 1.1382 0.9318 0.4292
1.0598 0.5731 400 1.0422 0.8139 0.3616
0.9698 0.7163 500 0.9831 0.8650 0.4124
1.0004 0.8596 600 0.9329 0.8296 0.3977
0.9246 1.0029 700 0.8990 0.7937 0.3763
0.8547 1.1461 800 0.8675 0.8520 0.4602
0.9163 1.2894 900 0.8364 0.7504 0.3477
0.8319 1.4327 1000 0.8205 0.7264 0.3456
0.8336 1.5759 1100 0.7872 0.7060 0.3151
0.7520 1.7192 1200 0.7683 0.7115 0.3259
0.7394 1.8625 1300 0.7534 0.6717 0.2963
0.6487 2.0057 1400 0.7362 0.6918 0.3321
0.6508 2.1490 1500 0.7349 0.6851 0.3106
0.6892 2.2923 1600 0.7207 0.7103 0.3391
0.6132 2.4355 1700 0.7182 0.6812 0.3050
0.6361 2.5788 1800 0.7013 0.7311 0.3462
0.6132 2.7221 1900 0.6935 0.6683 0.3133
0.6241 2.8653 2000 0.6861 0.7198 0.3492
0.5422 3.0086 2100 0.6782 0.6610 0.3011
0.5390 3.1519 2200 0.6831 0.6422 0.2983
0.5884 3.2951 2300 0.6734 0.6759 0.3239
0.5608 3.4384 2400 0.6720 0.6720 0.3242
0.5315 3.5817 2500 0.6606 0.6628 0.2952
0.4802 3.7249 2600 0.6539 0.6310 0.2849
0.5657 3.8682 2700 0.6458 0.6120 0.2736
0.4700 4.0115 2800 0.6510 0.6163 0.2805
0.4061 4.1547 2900 0.6600 0.6365 0.2918
0.4719 4.2980 3000 0.6517 0.6253 0.2820
0.4679 4.4413 3100 0.6493 0.5976 0.2680
0.4682 4.5845 3200 0.6440 0.6113 0.2746
0.4154 4.7278 3300 0.6364 0.6146 0.2859
0.5108 4.8711 3400 0.6324 0.5990 0.2680
0.3766 5.0143 3500 0.6392 0.5850 0.2659
0.3648 5.1576 3600 0.6478 0.6036 0.2768
0.3899 5.3009 3700 0.6429 0.6271 0.2914
0.4160 5.4441 3800 0.6422 0.6005 0.2725
0.4492 5.5874 3900 0.6426 0.6126 0.2843
0.3872 5.7307 4000 0.6377 0.5907 0.2726
0.4025 5.8739 4100 0.6374 0.5883 0.2673
0.2973 6.0172 4200 0.6476 0.5874 0.2697
0.2923 6.1605 4300 0.6602 0.6035 0.2777
0.2842 6.3037 4400 0.6654 0.5988 0.2764
0.3156 6.4470 4500 0.6530 0.6013 0.2767
0.3275 6.5903 4600 0.6513 0.6081 0.2814
0.2956 6.7335 4700 0.6502 0.5862 0.2713
0.2636 6.8768 4800 0.6507 0.5942 0.2769
0.2363 7.0201 4900 0.6660 0.5952 0.2780
0.2353 7.1633 5000 0.6786 0.6004 0.2811
0.2229 7.3066 5100 0.6801 0.6075 0.2845
0.2273 7.4499 5200 0.6838 0.5753 0.2628
0.2257 7.5931 5300 0.6772 0.5986 0.2773
0.2647 7.7364 5400 0.6763 0.5918 0.2764
0.2646 7.8797 5500 0.6764 0.5934 0.2731
0.1760 8.0229 5600 0.7018 0.5955 0.2724
0.1940 8.1662 5700 0.7091 0.5901 0.2689
0.1803 8.3095 5800 0.7106 0.6044 0.2804
0.1810 8.4527 5900 0.7054 0.5898 0.2679
0.1964 8.5960 6000 0.7074 0.6034 0.2812
0.2025 8.7393 6100 0.7128 0.5893 0.2706
0.1862 8.8825 6200 0.7079 0.5946 0.2733
0.1495 9.0258 6300 0.7216 0.5939 0.2731
0.1288 9.1691 6400 0.7314 0.5917 0.2742
0.1245 9.3123 6500 0.7353 0.5918 0.2757
0.1432 9.4556 6600 0.7376 0.5950 0.2741
0.1398 9.5989 6700 0.7387 0.5977 0.2800
0.1250 9.7421 6800 0.7390 0.6063 0.2859
0.1169 9.8854 6900 0.7369 0.6002 0.2809

Framework versions

  • PEFT 0.18.1
  • Transformers 5.3.0.dev0
  • Pytorch 2.10.0+cu128
  • Datasets 4.0.0
  • Tokenizers 0.22.2
Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sudoping01/karato-whisper-latin

Adapter
(322)
this model