File size: 2,715 Bytes
4c6aa59 5689e74 4c6aa59 05f48a0 4c6aa59 9ec5ce8 4c6aa59 9ec5ce8 4c6aa59 05f48a0 4c6aa59 9ec5ce8 d1c7a63 9ec5ce8 4c6aa59 5689e74 4c6aa59 5689e74 d1c7a63 4c6aa59 99f96ae 4c6aa59 99f96ae 4c6aa59 a30593b 99f96ae 4c6aa59 e778dac 9086a59 14f6058 a30593b 99f96ae 4c6aa59 e778dac 4c6aa59 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 |
---
license: apache-2.0
tags:
- google/fleurs
- generated_from_trainer
- automatic-speech-recognition
- pashto
- ps
datasets:
- fleurs
metrics:
- wer
base_model: facebook/wav2vec2-xls-r-300m
model-index:
- name: facebook/wav2vec2-xls-r-300m
results:
- task:
type: automatic-speech-recognition
name: Automatic Speech Recognition
dataset:
name: google/fleurs
type: google/fleurs
args: 'config: ps_af, split: test'
metrics:
- type: wer
value: 51.59447476125512
name: Wer
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# facebook/wav2vec2-xls-r-300m
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the GOOGLE/FLEURS - PS_AF dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9162
- Wer: 51.59
- Cer: 19.72
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7.5e-07
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- training_steps: 6000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Cer | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:------:|:---------------:|:------:|
| 5.0767 | 6.33 | 500 | 1.0 | 4.8783 | 1.0 |
| 3.1156 | 12.66 | 1000 | 1.0 | 3.0990 | 1.0 |
| 1.3506 | 18.99 | 1500 | 0.2889 | 1.1056 | 0.7031 |
| 0.9997 | 25.32 | 2000 | 0.2301 | 0.9191 | 0.5944 |
| 0.7838 | 31.65 | 2500 | 0.2152 | 0.8952 | 0.5556 |
| 0.6665 | 37.97 | 3000 | 0.2017 | 0.8908 | 0.5252 |
| 0.6265 | 44.3 | 3500 | 0.1954 | 0.9063 | 0.5133 |
| 0.5935 | 50.63 | 4000 | 0.1969 | 0.9162 | 0.5156 |
| 0.5174 | 56.96 | 4500 | 0.1972 | 0.9287 | 0.5140 |
| 0.5462 | 63.29 | 5000 | 0.1974 | 0.9370 | 0.5138 |
| 0.5564 | 69.62 | 5500 | 0.1977 | 0.9461 | 0.5148 |
| 0.5252 | 75.95 | 6000 | 0.9505 | 0.5118 | 0.1969 |
### Framework versions
- Transformers 4.26.0.dev0
- Pytorch 1.13.1+cu117
- Datasets 2.7.1.dev0
- Tokenizers 0.13.2
|