manuth commited on
Commit
0d5aa9b
·
verified ·
1 Parent(s): 9298795

End of training

Browse files
README.md CHANGED
@@ -8,9 +8,24 @@ tags:
8
  - generated_from_trainer
9
  datasets:
10
  - khmer-coupus
 
 
11
  model-index:
12
  - name: Whisper Large V3 Turbo Khmer
13
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -19,6 +34,9 @@ should probably proofread and complete it, then remove this comment. -->
19
  # Whisper Large V3 Turbo Khmer
20
 
21
  This model is a fine-tuned version of [openai/whisper-large-v3-turbo](https://huggingface.co/openai/whisper-large-v3-turbo) on the Common Voice 11.0 dataset.
 
 
 
22
 
23
  ## Model description
24
 
@@ -43,15 +61,25 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - gradient_accumulation_steps: 4
45
  - total_train_batch_size: 32
46
- - optimizer: Use OptimizerNames.ADAMW_HF with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 500
49
  - training_steps: 4000
50
  - mixed_precision_training: Native AMP
51
 
 
 
 
 
 
 
 
 
 
 
52
  ### Framework versions
53
 
54
  - Transformers 4.49.0
55
- - Pytorch 2.5.1+cu124
56
  - Datasets 3.3.1
57
  - Tokenizers 0.21.0
 
8
  - generated_from_trainer
9
  datasets:
10
  - khmer-coupus
11
+ metrics:
12
+ - wer
13
  model-index:
14
  - name: Whisper Large V3 Turbo Khmer
15
+ results:
16
+ - task:
17
+ name: Automatic Speech Recognition
18
+ type: automatic-speech-recognition
19
+ dataset:
20
+ name: Common Voice 11.0
21
+ type: khmer-coupus
22
+ config: km_kh
23
+ split: test
24
+ args: 'config: khm, split: test'
25
+ metrics:
26
+ - name: Wer
27
+ type: wer
28
+ value: 101.09561752988047
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
  # Whisper Large V3 Turbo Khmer
35
 
36
  This model is a fine-tuned version of [openai/whisper-large-v3-turbo](https://huggingface.co/openai/whisper-large-v3-turbo) on the Common Voice 11.0 dataset.
37
+ It achieves the following results on the evaluation set:
38
+ - Loss: 0.6277
39
+ - Wer: 101.0956
40
 
41
  ## Model description
42
 
 
61
  - seed: 42
62
  - gradient_accumulation_steps: 4
63
  - total_train_batch_size: 32
64
+ - optimizer: Use adamw_hf with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
65
  - lr_scheduler_type: linear
66
  - lr_scheduler_warmup_steps: 500
67
  - training_steps: 4000
68
  - mixed_precision_training: Native AMP
69
 
70
+ ### Training results
71
+
72
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
73
+ |:-------------:|:-------:|:----:|:---------------:|:--------:|
74
+ | 0.3899 | 15.8835 | 1000 | 0.6277 | 101.0956 |
75
+ | 0.0522 | 31.7550 | 2000 | 0.7151 | 102.6394 |
76
+ | 0.0051 | 47.6265 | 3000 | 1.0525 | 104.0090 |
77
+ | 0.0023 | 63.4980 | 4000 | 1.1451 | 104.5319 |
78
+
79
+
80
  ### Framework versions
81
 
82
  - Transformers 4.49.0
83
+ - Pytorch 2.6.0+cu124
84
  - Datasets 3.3.1
85
  - Tokenizers 0.21.0
generation_config.json CHANGED
@@ -2,27 +2,27 @@
2
  "alignment_heads": [
3
  [
4
  2,
5
- 4
6
  ],
7
  [
8
- 2,
9
- 11
10
  ],
11
  [
12
  3,
13
- 3
14
  ],
15
  [
16
  3,
17
- 6
18
  ],
19
  [
20
  3,
21
- 11
22
  ],
23
  [
24
  3,
25
- 14
26
  ]
27
  ],
28
  "begin_suppress_tokens": [
@@ -132,15 +132,14 @@
132
  "<|vi|>": 50278,
133
  "<|yi|>": 50335,
134
  "<|yo|>": 50325,
135
- "<|yue|>": 50358,
136
  "<|zh|>": 50260
137
  },
138
  "language": "khmer",
139
  "max_initial_timestamp_index": 50,
140
  "max_length": 448,
141
- "no_timestamps_token_id": 50364,
142
  "pad_token_id": 50257,
143
- "prev_sot_token_id": 50362,
144
  "return_timestamps": false,
145
  "suppress_tokens": [
146
  1,
@@ -226,16 +225,16 @@
226
  49870,
227
  50254,
228
  50258,
 
229
  50359,
230
  50360,
231
  50361,
232
- 50362,
233
- 50363
234
  ],
235
  "task": "transcribe",
236
  "task_to_id": {
237
- "transcribe": 50360,
238
- "translate": 50359
239
  },
240
  "transformers_version": "4.49.0"
241
  }
 
2
  "alignment_heads": [
3
  [
4
  2,
5
+ 2
6
  ],
7
  [
8
+ 3,
9
+ 0
10
  ],
11
  [
12
  3,
13
+ 2
14
  ],
15
  [
16
  3,
17
+ 3
18
  ],
19
  [
20
  3,
21
+ 4
22
  ],
23
  [
24
  3,
25
+ 5
26
  ]
27
  ],
28
  "begin_suppress_tokens": [
 
132
  "<|vi|>": 50278,
133
  "<|yi|>": 50335,
134
  "<|yo|>": 50325,
 
135
  "<|zh|>": 50260
136
  },
137
  "language": "khmer",
138
  "max_initial_timestamp_index": 50,
139
  "max_length": 448,
140
+ "no_timestamps_token_id": 50363,
141
  "pad_token_id": 50257,
142
+ "prev_sot_token_id": 50361,
143
  "return_timestamps": false,
144
  "suppress_tokens": [
145
  1,
 
225
  49870,
226
  50254,
227
  50258,
228
+ 50358,
229
  50359,
230
  50360,
231
  50361,
232
+ 50362
 
233
  ],
234
  "task": "transcribe",
235
  "task_to_id": {
236
+ "transcribe": 50359,
237
+ "translate": 50358
238
  },
239
  "transformers_version": "4.49.0"
240
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2b74b1a475161cd0a6ec5e9ffb1d35cefedcfd5c2ef00a564905ff32be1d248
3
  size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb02400f7c8ee414f095dfa9a72cbc5859f9ec813cffeed2862e30ac549fad3
3
  size 151061672
runs/Feb19_01-41-03_rathanak/events.out.tfevents.1739904067.rathanak.62045.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91657d73651d52739fea6cd30fc4ca633feb113e8cd1b321ca5fcbeed8bf0e20
3
- size 33089
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e25c5f38d162034fade99719d308ec7ece71ee56014011e2dceaad47ef9217a7
3
+ size 42201
runs/Feb19_01-41-03_rathanak/events.out.tfevents.1739917632.rathanak.62045.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdf0906581cee9c3987d85994a6f808a251d2ca7ce660a5a9001d844857642a4
3
+ size 406