yonathanstwn commited on
Commit
6311ad0
·
1 Parent(s): 924fa6a

End of training

Browse files
.gitattributes CHANGED
@@ -32,5 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
- last-checkpoint/tokenizer.json filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
35
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
last-checkpoint/generation_config.json → generation_config.json RENAMED
File without changes
last-checkpoint/config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "_name_or_path": "facebook/nllb-200-distilled-600M",
3
- "activation_dropout": 0.0,
4
- "activation_function": "relu",
5
- "architectures": [
6
- "M2M100ForConditionalGeneration"
7
- ],
8
- "attention_dropout": 0.1,
9
- "bos_token_id": 0,
10
- "d_model": 1024,
11
- "decoder_attention_heads": 16,
12
- "decoder_ffn_dim": 4096,
13
- "decoder_layerdrop": 0,
14
- "decoder_layers": 12,
15
- "decoder_start_token_id": 2,
16
- "dropout": 0.1,
17
- "encoder_attention_heads": 16,
18
- "encoder_ffn_dim": 4096,
19
- "encoder_layerdrop": 0,
20
- "encoder_layers": 12,
21
- "eos_token_id": 2,
22
- "init_std": 0.02,
23
- "is_encoder_decoder": true,
24
- "max_length": 200,
25
- "max_position_embeddings": 1024,
26
- "model_type": "m2m_100",
27
- "num_hidden_layers": 12,
28
- "pad_token_id": 1,
29
- "scale_embedding": true,
30
- "tokenizer_class": "NllbTokenizer",
31
- "torch_dtype": "float32",
32
- "transformers_version": "4.26.1",
33
- "use_cache": true,
34
- "vocab_size": 256206
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2ed4d6179a5d70ed99cf0fa7418fef2a20cb5560a2c4dcbb7dc4c46c446e7d5
3
- size 4920898947
 
 
 
 
last-checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b93f8641130a19f66c57f8ecf05dcf389906368a755f7f87bb668bde6600c5f6
3
- size 2468874377
 
 
 
 
last-checkpoint/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6dc3928bf65e95ad7412303bf594f62a1522c17b5d9e03e2d2a34b9029812b49
3
- size 14511
 
 
 
 
last-checkpoint/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:be3d6604c0bb0312bc94e5a6c7c044a2fb4b1ff121edd281d8f142b5f1ffce35
3
- size 627
 
 
 
 
last-checkpoint/sentencepiece.bpe.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:14bb8dfb35c0ffdea7bc01e56cea38b9e3d5efcdcb9c251d6b40538e1aab555a
3
- size 4852054
 
 
 
 
last-checkpoint/special_tokens_map.json DELETED
@@ -1,219 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "ace_Arab",
4
- "ace_Latn",
5
- "acm_Arab",
6
- "acq_Arab",
7
- "aeb_Arab",
8
- "afr_Latn",
9
- "ajp_Arab",
10
- "aka_Latn",
11
- "amh_Ethi",
12
- "apc_Arab",
13
- "arb_Arab",
14
- "ars_Arab",
15
- "ary_Arab",
16
- "arz_Arab",
17
- "asm_Beng",
18
- "ast_Latn",
19
- "awa_Deva",
20
- "ayr_Latn",
21
- "azb_Arab",
22
- "azj_Latn",
23
- "bak_Cyrl",
24
- "bam_Latn",
25
- "ban_Latn",
26
- "bel_Cyrl",
27
- "bem_Latn",
28
- "ben_Beng",
29
- "bho_Deva",
30
- "bjn_Arab",
31
- "bjn_Latn",
32
- "bod_Tibt",
33
- "bos_Latn",
34
- "bug_Latn",
35
- "bul_Cyrl",
36
- "cat_Latn",
37
- "ceb_Latn",
38
- "ces_Latn",
39
- "cjk_Latn",
40
- "ckb_Arab",
41
- "crh_Latn",
42
- "cym_Latn",
43
- "dan_Latn",
44
- "deu_Latn",
45
- "dik_Latn",
46
- "dyu_Latn",
47
- "dzo_Tibt",
48
- "ell_Grek",
49
- "eng_Latn",
50
- "epo_Latn",
51
- "est_Latn",
52
- "eus_Latn",
53
- "ewe_Latn",
54
- "fao_Latn",
55
- "pes_Arab",
56
- "fij_Latn",
57
- "fin_Latn",
58
- "fon_Latn",
59
- "fra_Latn",
60
- "fur_Latn",
61
- "fuv_Latn",
62
- "gla_Latn",
63
- "gle_Latn",
64
- "glg_Latn",
65
- "grn_Latn",
66
- "guj_Gujr",
67
- "hat_Latn",
68
- "hau_Latn",
69
- "heb_Hebr",
70
- "hin_Deva",
71
- "hne_Deva",
72
- "hrv_Latn",
73
- "hun_Latn",
74
- "hye_Armn",
75
- "ibo_Latn",
76
- "ilo_Latn",
77
- "ind_Latn",
78
- "isl_Latn",
79
- "ita_Latn",
80
- "jav_Latn",
81
- "jpn_Jpan",
82
- "kab_Latn",
83
- "kac_Latn",
84
- "kam_Latn",
85
- "kan_Knda",
86
- "kas_Arab",
87
- "kas_Deva",
88
- "kat_Geor",
89
- "knc_Arab",
90
- "knc_Latn",
91
- "kaz_Cyrl",
92
- "kbp_Latn",
93
- "kea_Latn",
94
- "khm_Khmr",
95
- "kik_Latn",
96
- "kin_Latn",
97
- "kir_Cyrl",
98
- "kmb_Latn",
99
- "kon_Latn",
100
- "kor_Hang",
101
- "kmr_Latn",
102
- "lao_Laoo",
103
- "lvs_Latn",
104
- "lij_Latn",
105
- "lim_Latn",
106
- "lin_Latn",
107
- "lit_Latn",
108
- "lmo_Latn",
109
- "ltg_Latn",
110
- "ltz_Latn",
111
- "lua_Latn",
112
- "lug_Latn",
113
- "luo_Latn",
114
- "lus_Latn",
115
- "mag_Deva",
116
- "mai_Deva",
117
- "mal_Mlym",
118
- "mar_Deva",
119
- "min_Latn",
120
- "mkd_Cyrl",
121
- "plt_Latn",
122
- "mlt_Latn",
123
- "mni_Beng",
124
- "khk_Cyrl",
125
- "mos_Latn",
126
- "mri_Latn",
127
- "zsm_Latn",
128
- "mya_Mymr",
129
- "nld_Latn",
130
- "nno_Latn",
131
- "nob_Latn",
132
- "npi_Deva",
133
- "nso_Latn",
134
- "nus_Latn",
135
- "nya_Latn",
136
- "oci_Latn",
137
- "gaz_Latn",
138
- "ory_Orya",
139
- "pag_Latn",
140
- "pan_Guru",
141
- "pap_Latn",
142
- "pol_Latn",
143
- "por_Latn",
144
- "prs_Arab",
145
- "pbt_Arab",
146
- "quy_Latn",
147
- "ron_Latn",
148
- "run_Latn",
149
- "rus_Cyrl",
150
- "sag_Latn",
151
- "san_Deva",
152
- "sat_Beng",
153
- "scn_Latn",
154
- "shn_Mymr",
155
- "sin_Sinh",
156
- "slk_Latn",
157
- "slv_Latn",
158
- "smo_Latn",
159
- "sna_Latn",
160
- "snd_Arab",
161
- "som_Latn",
162
- "sot_Latn",
163
- "spa_Latn",
164
- "als_Latn",
165
- "srd_Latn",
166
- "srp_Cyrl",
167
- "ssw_Latn",
168
- "sun_Latn",
169
- "swe_Latn",
170
- "swh_Latn",
171
- "szl_Latn",
172
- "tam_Taml",
173
- "tat_Cyrl",
174
- "tel_Telu",
175
- "tgk_Cyrl",
176
- "tgl_Latn",
177
- "tha_Thai",
178
- "tir_Ethi",
179
- "taq_Latn",
180
- "taq_Tfng",
181
- "tpi_Latn",
182
- "tsn_Latn",
183
- "tso_Latn",
184
- "tuk_Latn",
185
- "tum_Latn",
186
- "tur_Latn",
187
- "twi_Latn",
188
- "tzm_Tfng",
189
- "uig_Arab",
190
- "ukr_Cyrl",
191
- "umb_Latn",
192
- "urd_Arab",
193
- "uzn_Latn",
194
- "vec_Latn",
195
- "vie_Latn",
196
- "war_Latn",
197
- "wol_Latn",
198
- "xho_Latn",
199
- "ydd_Hebr",
200
- "yor_Latn",
201
- "yue_Hant",
202
- "zho_Hans",
203
- "zho_Hant",
204
- "zul_Latn"
205
- ],
206
- "bos_token": "<s>",
207
- "cls_token": "<s>",
208
- "eos_token": "</s>",
209
- "mask_token": {
210
- "content": "<mask>",
211
- "lstrip": true,
212
- "normalized": true,
213
- "rstrip": false,
214
- "single_word": false
215
- },
216
- "pad_token": "<pad>",
217
- "sep_token": "</s>",
218
- "unk_token": "<unk>"
219
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dec87343dd7b147b84072891b984b69471e9b7539eee4e544a0450ffc88e0e8
3
- size 17331294
 
 
 
 
last-checkpoint/tokenizer_config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "additional_special_tokens": null,
3
- "bos_token": "<s>",
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "mask_token": {
7
- "__type": "AddedToken",
8
- "content": "<mask>",
9
- "lstrip": true,
10
- "normalized": true,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "model_max_length": 1024,
15
- "name_or_path": "facebook/nllb-200-distilled-600M",
16
- "pad_token": "<pad>",
17
- "sep_token": "</s>",
18
- "sp_model_kwargs": {},
19
- "special_tokens_map_file": null,
20
- "src_lang": "eng_Latn",
21
- "tgt_lang": "ind_Latn",
22
- "tokenizer_class": "NllbTokenizer",
23
- "unk_token": "<unk>"
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/trainer_state.json DELETED
@@ -1,166 +0,0 @@
1
- {
2
- "best_metric": 0.47876495122909546,
3
- "best_model_checkpoint": "models/nllb-en-id-ccmatrix/checkpoint-253125",
4
- "epoch": 10.0,
5
- "global_step": 281250,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 1.0,
12
- "learning_rate": 9.129846708746619e-06,
13
- "loss": 0.606,
14
- "step": 28125
15
- },
16
- {
17
- "epoch": 1.0,
18
- "eval_bleu": 64.1268,
19
- "eval_loss": 0.524931013584137,
20
- "eval_runtime": 3349.5808,
21
- "eval_samples_per_second": 29.854,
22
- "eval_steps_per_second": 0.933,
23
- "step": 28125
24
- },
25
- {
26
- "epoch": 2.0,
27
- "learning_rate": 8.115419296663661e-06,
28
- "loss": 0.4943,
29
- "step": 56250
30
- },
31
- {
32
- "epoch": 2.0,
33
- "eval_bleu": 64.7892,
34
- "eval_loss": 0.5042685866355896,
35
- "eval_runtime": 3375.6685,
36
- "eval_samples_per_second": 29.624,
37
- "eval_steps_per_second": 0.926,
38
- "step": 56250
39
- },
40
- {
41
- "epoch": 3.0,
42
- "learning_rate": 7.100991884580704e-06,
43
- "loss": 0.467,
44
- "step": 84375
45
- },
46
- {
47
- "epoch": 3.0,
48
- "eval_bleu": 65.2331,
49
- "eval_loss": 0.49452316761016846,
50
- "eval_runtime": 3358.2539,
51
- "eval_samples_per_second": 29.777,
52
- "eval_steps_per_second": 0.931,
53
- "step": 84375
54
- },
55
- {
56
- "epoch": 4.0,
57
- "learning_rate": 6.0865644724977455e-06,
58
- "loss": 0.4487,
59
- "step": 112500
60
- },
61
- {
62
- "epoch": 4.0,
63
- "eval_bleu": 65.5512,
64
- "eval_loss": 0.488678514957428,
65
- "eval_runtime": 3355.1997,
66
- "eval_samples_per_second": 29.804,
67
- "eval_steps_per_second": 0.931,
68
- "step": 112500
69
- },
70
- {
71
- "epoch": 5.0,
72
- "learning_rate": 5.072137060414789e-06,
73
- "loss": 0.4349,
74
- "step": 140625
75
- },
76
- {
77
- "epoch": 5.0,
78
- "eval_bleu": 65.6806,
79
- "eval_loss": 0.4842682480812073,
80
- "eval_runtime": 3377.8574,
81
- "eval_samples_per_second": 29.605,
82
- "eval_steps_per_second": 0.925,
83
- "step": 140625
84
- },
85
- {
86
- "epoch": 6.0,
87
- "learning_rate": 4.057709648331831e-06,
88
- "loss": 0.4242,
89
- "step": 168750
90
- },
91
- {
92
- "epoch": 6.0,
93
- "eval_bleu": 65.7774,
94
- "eval_loss": 0.48221057653427124,
95
- "eval_runtime": 3362.6872,
96
- "eval_samples_per_second": 29.738,
97
- "eval_steps_per_second": 0.929,
98
- "step": 168750
99
- },
100
- {
101
- "epoch": 7.0,
102
- "learning_rate": 3.0432822362488728e-06,
103
- "loss": 0.416,
104
- "step": 196875
105
- },
106
- {
107
- "epoch": 7.0,
108
- "eval_bleu": 65.8541,
109
- "eval_loss": 0.48012077808380127,
110
- "eval_runtime": 3379.151,
111
- "eval_samples_per_second": 29.593,
112
- "eval_steps_per_second": 0.925,
113
- "step": 196875
114
- },
115
- {
116
- "epoch": 8.0,
117
- "learning_rate": 2.0288548241659153e-06,
118
- "loss": 0.4098,
119
- "step": 225000
120
- },
121
- {
122
- "epoch": 8.0,
123
- "eval_bleu": 65.9652,
124
- "eval_loss": 0.4800304174423218,
125
- "eval_runtime": 3370.7435,
126
- "eval_samples_per_second": 29.667,
127
- "eval_steps_per_second": 0.927,
128
- "step": 225000
129
- },
130
- {
131
- "epoch": 9.0,
132
- "learning_rate": 1.0144274120829577e-06,
133
- "loss": 0.4052,
134
- "step": 253125
135
- },
136
- {
137
- "epoch": 9.0,
138
- "eval_bleu": 65.9701,
139
- "eval_loss": 0.47876495122909546,
140
- "eval_runtime": 3378.6802,
141
- "eval_samples_per_second": 29.597,
142
- "eval_steps_per_second": 0.925,
143
- "step": 253125
144
- },
145
- {
146
- "epoch": 10.0,
147
- "learning_rate": 0.0,
148
- "loss": 0.4023,
149
- "step": 281250
150
- },
151
- {
152
- "epoch": 10.0,
153
- "eval_bleu": 65.9837,
154
- "eval_loss": 0.47907769680023193,
155
- "eval_runtime": 3373.8692,
156
- "eval_samples_per_second": 29.64,
157
- "eval_steps_per_second": 0.926,
158
- "step": 281250
159
- }
160
- ],
161
- "max_steps": 281250,
162
- "num_train_epochs": 10,
163
- "total_flos": 7.671437872925245e+17,
164
- "trial_name": null,
165
- "trial_params": null
166
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ce8a25a4e3959df6ac5c042f17d7dbec4233da4feb7c11b467ccefc47e5ab97
3
- size 3643
 
 
 
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b93f8641130a19f66c57f8ecf05dcf389906368a755f7f87bb668bde6600c5f6
3
  size 2468874377
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d05009dbe23b1d997be5391df7f03167d76ffaee130c7dddc471568d4326068
3
  size 2468874377