wav2vec2-bloom-speech-kan
Model description
- Homepage: SIL AI
- Point of Contact: SIL AI email
- Source Data: Bloom Library
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the SIL-AI/bloom-speech - KAN (Kannada) dataset. It achieves the following results on the evaluation set:
- Loss: 0.9781
- Wer: 0.5096
- Cer: 0.1228
Users should refer to the original model for tutorials on using a trained model for inference.
Intended uses & limitations
Users of this model must abide by the SIL RAIL-M License.
This model is created as a proof of concept and no guarantees are made regarding the performance of the model is specific situations.
Training and evaluation data
Training, Validation, and Test datasets were generated from the same corpus, ensuring that no duplicate files were used.
Training procedure
Standard finetuning of XLS-R was used based on the examples in the Hugging Face Transformers Github
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 250
- num_epochs: 1000.0
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
---|---|---|---|---|---|
No log | 35.67 | 250 | 3.6338 | 1.0160 | 0.8984 |
4.5691 | 71.4 | 500 | 0.9781 | 0.6699 | 0.1594 |
4.5691 | 107.13 | 750 | 0.8616 | 0.6282 | 0.1534 |
0.0707 | 142.8 | 1000 | 0.9301 | 0.5833 | 0.1432 |
0.0707 | 178.53 | 1250 | 0.9178 | 0.5513 | 0.1275 |
0.0321 | 214.27 | 1500 | 0.9461 | 0.5929 | 0.1398 |
0.0321 | 249.93 | 1750 | 1.0137 | 0.5481 | 0.1292 |
0.0202 | 285.67 | 2000 | 0.9781 | 0.5096 | 0.1228 |
0.0202 | 321.4 | 2250 | 0.9078 | 0.5128 | 0.1237 |
0.0114 | 357.13 | 2500 | 0.9275 | 0.5160 | 0.1241 |
0.0114 | 392.8 | 2750 | 0.9977 | 0.5128 | 0.1194 |
Framework versions
- Transformers 4.21.0.dev0
- Pytorch 1.9.0+cu111
- Datasets 2.2.2
- Tokenizers 0.12.1