adrianSauer commited on
Commit
a63a0b1
·
verified ·
1 Parent(s): b0a9ff3

End of training

Browse files
Files changed (1) hide show
  1. README.md +13 -11
README.md CHANGED
@@ -19,8 +19,8 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [glob-asr/wav2vec2-large-xls-r-300m-guarani-small](https://huggingface.co/glob-asr/wav2vec2-large-xls-r-300m-guarani-small) on the Common Voice 16 dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.3202
23
- - Cer: 7.2954
24
 
25
  ## Model description
26
 
@@ -40,9 +40,11 @@ More information needed
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 1e-05
43
- - train_batch_size: 16
44
  - eval_batch_size: 16
45
  - seed: 42
 
 
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant_with_warmup
48
  - lr_scheduler_warmup_steps: 50
@@ -53,16 +55,16 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Cer |
55
  |:-------------:|:------:|:----:|:---------------:|:------:|
56
- | 0.4174 | 1.0101 | 100 | 0.3535 | 8.1385 |
57
- | 0.3411 | 2.0202 | 200 | 0.3387 | 7.8574 |
58
- | 0.2905 | 3.0303 | 300 | 0.3278 | 7.6076 |
59
- | 0.2591 | 4.0404 | 400 | 0.3214 | 7.3734 |
60
- | 0.251 | 5.0505 | 500 | 0.3202 | 7.2954 |
61
 
62
 
63
  ### Framework versions
64
 
65
- - Transformers 4.40.1
66
- - Pytorch 2.2.1+cu121
67
- - Datasets 2.19.0
68
  - Tokenizers 0.19.1
 
19
 
20
  This model is a fine-tuned version of [glob-asr/wav2vec2-large-xls-r-300m-guarani-small](https://huggingface.co/glob-asr/wav2vec2-large-xls-r-300m-guarani-small) on the Common Voice 16 dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.3490
23
+ - Cer: 7.9739
24
 
25
  ## Model description
26
 
 
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 1e-05
43
+ - train_batch_size: 8
44
  - eval_batch_size: 16
45
  - seed: 42
46
+ - gradient_accumulation_steps: 2
47
+ - total_train_batch_size: 16
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: constant_with_warmup
50
  - lr_scheduler_warmup_steps: 50
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Cer |
57
  |:-------------:|:------:|:----:|:---------------:|:------:|
58
+ | 0.4171 | 1.0152 | 100 | 0.3798 | 8.5448 |
59
+ | 0.3367 | 2.0305 | 200 | 0.3633 | 8.4451 |
60
+ | 0.2914 | 3.0457 | 300 | 0.3564 | 8.1279 |
61
+ | 0.261 | 4.0609 | 400 | 0.3494 | 7.9014 |
62
+ | 0.2464 | 5.0761 | 500 | 0.3490 | 7.9739 |
63
 
64
 
65
  ### Framework versions
66
 
67
+ - Transformers 4.44.0
68
+ - Pytorch 2.3.1+cu121
69
+ - Datasets 2.21.0
70
  - Tokenizers 0.19.1