Commit
·
6311ad0
1
Parent(s):
924fa6a
End of training
Browse files- .gitattributes +0 -1
- last-checkpoint/generation_config.json → generation_config.json +0 -0
- last-checkpoint/config.json +0 -35
- last-checkpoint/optimizer.pt +0 -3
- last-checkpoint/pytorch_model.bin +0 -3
- last-checkpoint/rng_state.pth +0 -3
- last-checkpoint/scheduler.pt +0 -3
- last-checkpoint/sentencepiece.bpe.model +0 -3
- last-checkpoint/special_tokens_map.json +0 -219
- last-checkpoint/tokenizer.json +0 -3
- last-checkpoint/tokenizer_config.json +0 -24
- last-checkpoint/trainer_state.json +0 -166
- last-checkpoint/training_args.bin +0 -3
- pytorch_model.bin +1 -1
.gitattributes
CHANGED
@@ -32,5 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
-
last-checkpoint/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
36 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
35 |
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
last-checkpoint/generation_config.json → generation_config.json
RENAMED
File without changes
|
last-checkpoint/config.json
DELETED
@@ -1,35 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "facebook/nllb-200-distilled-600M",
|
3 |
-
"activation_dropout": 0.0,
|
4 |
-
"activation_function": "relu",
|
5 |
-
"architectures": [
|
6 |
-
"M2M100ForConditionalGeneration"
|
7 |
-
],
|
8 |
-
"attention_dropout": 0.1,
|
9 |
-
"bos_token_id": 0,
|
10 |
-
"d_model": 1024,
|
11 |
-
"decoder_attention_heads": 16,
|
12 |
-
"decoder_ffn_dim": 4096,
|
13 |
-
"decoder_layerdrop": 0,
|
14 |
-
"decoder_layers": 12,
|
15 |
-
"decoder_start_token_id": 2,
|
16 |
-
"dropout": 0.1,
|
17 |
-
"encoder_attention_heads": 16,
|
18 |
-
"encoder_ffn_dim": 4096,
|
19 |
-
"encoder_layerdrop": 0,
|
20 |
-
"encoder_layers": 12,
|
21 |
-
"eos_token_id": 2,
|
22 |
-
"init_std": 0.02,
|
23 |
-
"is_encoder_decoder": true,
|
24 |
-
"max_length": 200,
|
25 |
-
"max_position_embeddings": 1024,
|
26 |
-
"model_type": "m2m_100",
|
27 |
-
"num_hidden_layers": 12,
|
28 |
-
"pad_token_id": 1,
|
29 |
-
"scale_embedding": true,
|
30 |
-
"tokenizer_class": "NllbTokenizer",
|
31 |
-
"torch_dtype": "float32",
|
32 |
-
"transformers_version": "4.26.1",
|
33 |
-
"use_cache": true,
|
34 |
-
"vocab_size": 256206
|
35 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b2ed4d6179a5d70ed99cf0fa7418fef2a20cb5560a2c4dcbb7dc4c46c446e7d5
|
3 |
-
size 4920898947
|
|
|
|
|
|
|
|
last-checkpoint/pytorch_model.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b93f8641130a19f66c57f8ecf05dcf389906368a755f7f87bb668bde6600c5f6
|
3 |
-
size 2468874377
|
|
|
|
|
|
|
|
last-checkpoint/rng_state.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:6dc3928bf65e95ad7412303bf594f62a1522c17b5d9e03e2d2a34b9029812b49
|
3 |
-
size 14511
|
|
|
|
|
|
|
|
last-checkpoint/scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:be3d6604c0bb0312bc94e5a6c7c044a2fb4b1ff121edd281d8f142b5f1ffce35
|
3 |
-
size 627
|
|
|
|
|
|
|
|
last-checkpoint/sentencepiece.bpe.model
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:14bb8dfb35c0ffdea7bc01e56cea38b9e3d5efcdcb9c251d6b40538e1aab555a
|
3 |
-
size 4852054
|
|
|
|
|
|
|
|
last-checkpoint/special_tokens_map.json
DELETED
@@ -1,219 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"additional_special_tokens": [
|
3 |
-
"ace_Arab",
|
4 |
-
"ace_Latn",
|
5 |
-
"acm_Arab",
|
6 |
-
"acq_Arab",
|
7 |
-
"aeb_Arab",
|
8 |
-
"afr_Latn",
|
9 |
-
"ajp_Arab",
|
10 |
-
"aka_Latn",
|
11 |
-
"amh_Ethi",
|
12 |
-
"apc_Arab",
|
13 |
-
"arb_Arab",
|
14 |
-
"ars_Arab",
|
15 |
-
"ary_Arab",
|
16 |
-
"arz_Arab",
|
17 |
-
"asm_Beng",
|
18 |
-
"ast_Latn",
|
19 |
-
"awa_Deva",
|
20 |
-
"ayr_Latn",
|
21 |
-
"azb_Arab",
|
22 |
-
"azj_Latn",
|
23 |
-
"bak_Cyrl",
|
24 |
-
"bam_Latn",
|
25 |
-
"ban_Latn",
|
26 |
-
"bel_Cyrl",
|
27 |
-
"bem_Latn",
|
28 |
-
"ben_Beng",
|
29 |
-
"bho_Deva",
|
30 |
-
"bjn_Arab",
|
31 |
-
"bjn_Latn",
|
32 |
-
"bod_Tibt",
|
33 |
-
"bos_Latn",
|
34 |
-
"bug_Latn",
|
35 |
-
"bul_Cyrl",
|
36 |
-
"cat_Latn",
|
37 |
-
"ceb_Latn",
|
38 |
-
"ces_Latn",
|
39 |
-
"cjk_Latn",
|
40 |
-
"ckb_Arab",
|
41 |
-
"crh_Latn",
|
42 |
-
"cym_Latn",
|
43 |
-
"dan_Latn",
|
44 |
-
"deu_Latn",
|
45 |
-
"dik_Latn",
|
46 |
-
"dyu_Latn",
|
47 |
-
"dzo_Tibt",
|
48 |
-
"ell_Grek",
|
49 |
-
"eng_Latn",
|
50 |
-
"epo_Latn",
|
51 |
-
"est_Latn",
|
52 |
-
"eus_Latn",
|
53 |
-
"ewe_Latn",
|
54 |
-
"fao_Latn",
|
55 |
-
"pes_Arab",
|
56 |
-
"fij_Latn",
|
57 |
-
"fin_Latn",
|
58 |
-
"fon_Latn",
|
59 |
-
"fra_Latn",
|
60 |
-
"fur_Latn",
|
61 |
-
"fuv_Latn",
|
62 |
-
"gla_Latn",
|
63 |
-
"gle_Latn",
|
64 |
-
"glg_Latn",
|
65 |
-
"grn_Latn",
|
66 |
-
"guj_Gujr",
|
67 |
-
"hat_Latn",
|
68 |
-
"hau_Latn",
|
69 |
-
"heb_Hebr",
|
70 |
-
"hin_Deva",
|
71 |
-
"hne_Deva",
|
72 |
-
"hrv_Latn",
|
73 |
-
"hun_Latn",
|
74 |
-
"hye_Armn",
|
75 |
-
"ibo_Latn",
|
76 |
-
"ilo_Latn",
|
77 |
-
"ind_Latn",
|
78 |
-
"isl_Latn",
|
79 |
-
"ita_Latn",
|
80 |
-
"jav_Latn",
|
81 |
-
"jpn_Jpan",
|
82 |
-
"kab_Latn",
|
83 |
-
"kac_Latn",
|
84 |
-
"kam_Latn",
|
85 |
-
"kan_Knda",
|
86 |
-
"kas_Arab",
|
87 |
-
"kas_Deva",
|
88 |
-
"kat_Geor",
|
89 |
-
"knc_Arab",
|
90 |
-
"knc_Latn",
|
91 |
-
"kaz_Cyrl",
|
92 |
-
"kbp_Latn",
|
93 |
-
"kea_Latn",
|
94 |
-
"khm_Khmr",
|
95 |
-
"kik_Latn",
|
96 |
-
"kin_Latn",
|
97 |
-
"kir_Cyrl",
|
98 |
-
"kmb_Latn",
|
99 |
-
"kon_Latn",
|
100 |
-
"kor_Hang",
|
101 |
-
"kmr_Latn",
|
102 |
-
"lao_Laoo",
|
103 |
-
"lvs_Latn",
|
104 |
-
"lij_Latn",
|
105 |
-
"lim_Latn",
|
106 |
-
"lin_Latn",
|
107 |
-
"lit_Latn",
|
108 |
-
"lmo_Latn",
|
109 |
-
"ltg_Latn",
|
110 |
-
"ltz_Latn",
|
111 |
-
"lua_Latn",
|
112 |
-
"lug_Latn",
|
113 |
-
"luo_Latn",
|
114 |
-
"lus_Latn",
|
115 |
-
"mag_Deva",
|
116 |
-
"mai_Deva",
|
117 |
-
"mal_Mlym",
|
118 |
-
"mar_Deva",
|
119 |
-
"min_Latn",
|
120 |
-
"mkd_Cyrl",
|
121 |
-
"plt_Latn",
|
122 |
-
"mlt_Latn",
|
123 |
-
"mni_Beng",
|
124 |
-
"khk_Cyrl",
|
125 |
-
"mos_Latn",
|
126 |
-
"mri_Latn",
|
127 |
-
"zsm_Latn",
|
128 |
-
"mya_Mymr",
|
129 |
-
"nld_Latn",
|
130 |
-
"nno_Latn",
|
131 |
-
"nob_Latn",
|
132 |
-
"npi_Deva",
|
133 |
-
"nso_Latn",
|
134 |
-
"nus_Latn",
|
135 |
-
"nya_Latn",
|
136 |
-
"oci_Latn",
|
137 |
-
"gaz_Latn",
|
138 |
-
"ory_Orya",
|
139 |
-
"pag_Latn",
|
140 |
-
"pan_Guru",
|
141 |
-
"pap_Latn",
|
142 |
-
"pol_Latn",
|
143 |
-
"por_Latn",
|
144 |
-
"prs_Arab",
|
145 |
-
"pbt_Arab",
|
146 |
-
"quy_Latn",
|
147 |
-
"ron_Latn",
|
148 |
-
"run_Latn",
|
149 |
-
"rus_Cyrl",
|
150 |
-
"sag_Latn",
|
151 |
-
"san_Deva",
|
152 |
-
"sat_Beng",
|
153 |
-
"scn_Latn",
|
154 |
-
"shn_Mymr",
|
155 |
-
"sin_Sinh",
|
156 |
-
"slk_Latn",
|
157 |
-
"slv_Latn",
|
158 |
-
"smo_Latn",
|
159 |
-
"sna_Latn",
|
160 |
-
"snd_Arab",
|
161 |
-
"som_Latn",
|
162 |
-
"sot_Latn",
|
163 |
-
"spa_Latn",
|
164 |
-
"als_Latn",
|
165 |
-
"srd_Latn",
|
166 |
-
"srp_Cyrl",
|
167 |
-
"ssw_Latn",
|
168 |
-
"sun_Latn",
|
169 |
-
"swe_Latn",
|
170 |
-
"swh_Latn",
|
171 |
-
"szl_Latn",
|
172 |
-
"tam_Taml",
|
173 |
-
"tat_Cyrl",
|
174 |
-
"tel_Telu",
|
175 |
-
"tgk_Cyrl",
|
176 |
-
"tgl_Latn",
|
177 |
-
"tha_Thai",
|
178 |
-
"tir_Ethi",
|
179 |
-
"taq_Latn",
|
180 |
-
"taq_Tfng",
|
181 |
-
"tpi_Latn",
|
182 |
-
"tsn_Latn",
|
183 |
-
"tso_Latn",
|
184 |
-
"tuk_Latn",
|
185 |
-
"tum_Latn",
|
186 |
-
"tur_Latn",
|
187 |
-
"twi_Latn",
|
188 |
-
"tzm_Tfng",
|
189 |
-
"uig_Arab",
|
190 |
-
"ukr_Cyrl",
|
191 |
-
"umb_Latn",
|
192 |
-
"urd_Arab",
|
193 |
-
"uzn_Latn",
|
194 |
-
"vec_Latn",
|
195 |
-
"vie_Latn",
|
196 |
-
"war_Latn",
|
197 |
-
"wol_Latn",
|
198 |
-
"xho_Latn",
|
199 |
-
"ydd_Hebr",
|
200 |
-
"yor_Latn",
|
201 |
-
"yue_Hant",
|
202 |
-
"zho_Hans",
|
203 |
-
"zho_Hant",
|
204 |
-
"zul_Latn"
|
205 |
-
],
|
206 |
-
"bos_token": "<s>",
|
207 |
-
"cls_token": "<s>",
|
208 |
-
"eos_token": "</s>",
|
209 |
-
"mask_token": {
|
210 |
-
"content": "<mask>",
|
211 |
-
"lstrip": true,
|
212 |
-
"normalized": true,
|
213 |
-
"rstrip": false,
|
214 |
-
"single_word": false
|
215 |
-
},
|
216 |
-
"pad_token": "<pad>",
|
217 |
-
"sep_token": "</s>",
|
218 |
-
"unk_token": "<unk>"
|
219 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/tokenizer.json
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:8dec87343dd7b147b84072891b984b69471e9b7539eee4e544a0450ffc88e0e8
|
3 |
-
size 17331294
|
|
|
|
|
|
|
|
last-checkpoint/tokenizer_config.json
DELETED
@@ -1,24 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"additional_special_tokens": null,
|
3 |
-
"bos_token": "<s>",
|
4 |
-
"cls_token": "<s>",
|
5 |
-
"eos_token": "</s>",
|
6 |
-
"mask_token": {
|
7 |
-
"__type": "AddedToken",
|
8 |
-
"content": "<mask>",
|
9 |
-
"lstrip": true,
|
10 |
-
"normalized": true,
|
11 |
-
"rstrip": false,
|
12 |
-
"single_word": false
|
13 |
-
},
|
14 |
-
"model_max_length": 1024,
|
15 |
-
"name_or_path": "facebook/nllb-200-distilled-600M",
|
16 |
-
"pad_token": "<pad>",
|
17 |
-
"sep_token": "</s>",
|
18 |
-
"sp_model_kwargs": {},
|
19 |
-
"special_tokens_map_file": null,
|
20 |
-
"src_lang": "eng_Latn",
|
21 |
-
"tgt_lang": "ind_Latn",
|
22 |
-
"tokenizer_class": "NllbTokenizer",
|
23 |
-
"unk_token": "<unk>"
|
24 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/trainer_state.json
DELETED
@@ -1,166 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"best_metric": 0.47876495122909546,
|
3 |
-
"best_model_checkpoint": "models/nllb-en-id-ccmatrix/checkpoint-253125",
|
4 |
-
"epoch": 10.0,
|
5 |
-
"global_step": 281250,
|
6 |
-
"is_hyper_param_search": false,
|
7 |
-
"is_local_process_zero": true,
|
8 |
-
"is_world_process_zero": true,
|
9 |
-
"log_history": [
|
10 |
-
{
|
11 |
-
"epoch": 1.0,
|
12 |
-
"learning_rate": 9.129846708746619e-06,
|
13 |
-
"loss": 0.606,
|
14 |
-
"step": 28125
|
15 |
-
},
|
16 |
-
{
|
17 |
-
"epoch": 1.0,
|
18 |
-
"eval_bleu": 64.1268,
|
19 |
-
"eval_loss": 0.524931013584137,
|
20 |
-
"eval_runtime": 3349.5808,
|
21 |
-
"eval_samples_per_second": 29.854,
|
22 |
-
"eval_steps_per_second": 0.933,
|
23 |
-
"step": 28125
|
24 |
-
},
|
25 |
-
{
|
26 |
-
"epoch": 2.0,
|
27 |
-
"learning_rate": 8.115419296663661e-06,
|
28 |
-
"loss": 0.4943,
|
29 |
-
"step": 56250
|
30 |
-
},
|
31 |
-
{
|
32 |
-
"epoch": 2.0,
|
33 |
-
"eval_bleu": 64.7892,
|
34 |
-
"eval_loss": 0.5042685866355896,
|
35 |
-
"eval_runtime": 3375.6685,
|
36 |
-
"eval_samples_per_second": 29.624,
|
37 |
-
"eval_steps_per_second": 0.926,
|
38 |
-
"step": 56250
|
39 |
-
},
|
40 |
-
{
|
41 |
-
"epoch": 3.0,
|
42 |
-
"learning_rate": 7.100991884580704e-06,
|
43 |
-
"loss": 0.467,
|
44 |
-
"step": 84375
|
45 |
-
},
|
46 |
-
{
|
47 |
-
"epoch": 3.0,
|
48 |
-
"eval_bleu": 65.2331,
|
49 |
-
"eval_loss": 0.49452316761016846,
|
50 |
-
"eval_runtime": 3358.2539,
|
51 |
-
"eval_samples_per_second": 29.777,
|
52 |
-
"eval_steps_per_second": 0.931,
|
53 |
-
"step": 84375
|
54 |
-
},
|
55 |
-
{
|
56 |
-
"epoch": 4.0,
|
57 |
-
"learning_rate": 6.0865644724977455e-06,
|
58 |
-
"loss": 0.4487,
|
59 |
-
"step": 112500
|
60 |
-
},
|
61 |
-
{
|
62 |
-
"epoch": 4.0,
|
63 |
-
"eval_bleu": 65.5512,
|
64 |
-
"eval_loss": 0.488678514957428,
|
65 |
-
"eval_runtime": 3355.1997,
|
66 |
-
"eval_samples_per_second": 29.804,
|
67 |
-
"eval_steps_per_second": 0.931,
|
68 |
-
"step": 112500
|
69 |
-
},
|
70 |
-
{
|
71 |
-
"epoch": 5.0,
|
72 |
-
"learning_rate": 5.072137060414789e-06,
|
73 |
-
"loss": 0.4349,
|
74 |
-
"step": 140625
|
75 |
-
},
|
76 |
-
{
|
77 |
-
"epoch": 5.0,
|
78 |
-
"eval_bleu": 65.6806,
|
79 |
-
"eval_loss": 0.4842682480812073,
|
80 |
-
"eval_runtime": 3377.8574,
|
81 |
-
"eval_samples_per_second": 29.605,
|
82 |
-
"eval_steps_per_second": 0.925,
|
83 |
-
"step": 140625
|
84 |
-
},
|
85 |
-
{
|
86 |
-
"epoch": 6.0,
|
87 |
-
"learning_rate": 4.057709648331831e-06,
|
88 |
-
"loss": 0.4242,
|
89 |
-
"step": 168750
|
90 |
-
},
|
91 |
-
{
|
92 |
-
"epoch": 6.0,
|
93 |
-
"eval_bleu": 65.7774,
|
94 |
-
"eval_loss": 0.48221057653427124,
|
95 |
-
"eval_runtime": 3362.6872,
|
96 |
-
"eval_samples_per_second": 29.738,
|
97 |
-
"eval_steps_per_second": 0.929,
|
98 |
-
"step": 168750
|
99 |
-
},
|
100 |
-
{
|
101 |
-
"epoch": 7.0,
|
102 |
-
"learning_rate": 3.0432822362488728e-06,
|
103 |
-
"loss": 0.416,
|
104 |
-
"step": 196875
|
105 |
-
},
|
106 |
-
{
|
107 |
-
"epoch": 7.0,
|
108 |
-
"eval_bleu": 65.8541,
|
109 |
-
"eval_loss": 0.48012077808380127,
|
110 |
-
"eval_runtime": 3379.151,
|
111 |
-
"eval_samples_per_second": 29.593,
|
112 |
-
"eval_steps_per_second": 0.925,
|
113 |
-
"step": 196875
|
114 |
-
},
|
115 |
-
{
|
116 |
-
"epoch": 8.0,
|
117 |
-
"learning_rate": 2.0288548241659153e-06,
|
118 |
-
"loss": 0.4098,
|
119 |
-
"step": 225000
|
120 |
-
},
|
121 |
-
{
|
122 |
-
"epoch": 8.0,
|
123 |
-
"eval_bleu": 65.9652,
|
124 |
-
"eval_loss": 0.4800304174423218,
|
125 |
-
"eval_runtime": 3370.7435,
|
126 |
-
"eval_samples_per_second": 29.667,
|
127 |
-
"eval_steps_per_second": 0.927,
|
128 |
-
"step": 225000
|
129 |
-
},
|
130 |
-
{
|
131 |
-
"epoch": 9.0,
|
132 |
-
"learning_rate": 1.0144274120829577e-06,
|
133 |
-
"loss": 0.4052,
|
134 |
-
"step": 253125
|
135 |
-
},
|
136 |
-
{
|
137 |
-
"epoch": 9.0,
|
138 |
-
"eval_bleu": 65.9701,
|
139 |
-
"eval_loss": 0.47876495122909546,
|
140 |
-
"eval_runtime": 3378.6802,
|
141 |
-
"eval_samples_per_second": 29.597,
|
142 |
-
"eval_steps_per_second": 0.925,
|
143 |
-
"step": 253125
|
144 |
-
},
|
145 |
-
{
|
146 |
-
"epoch": 10.0,
|
147 |
-
"learning_rate": 0.0,
|
148 |
-
"loss": 0.4023,
|
149 |
-
"step": 281250
|
150 |
-
},
|
151 |
-
{
|
152 |
-
"epoch": 10.0,
|
153 |
-
"eval_bleu": 65.9837,
|
154 |
-
"eval_loss": 0.47907769680023193,
|
155 |
-
"eval_runtime": 3373.8692,
|
156 |
-
"eval_samples_per_second": 29.64,
|
157 |
-
"eval_steps_per_second": 0.926,
|
158 |
-
"step": 281250
|
159 |
-
}
|
160 |
-
],
|
161 |
-
"max_steps": 281250,
|
162 |
-
"num_train_epochs": 10,
|
163 |
-
"total_flos": 7.671437872925245e+17,
|
164 |
-
"trial_name": null,
|
165 |
-
"trial_params": null
|
166 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
last-checkpoint/training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:3ce8a25a4e3959df6ac5c042f17d7dbec4233da4feb7c11b467ccefc47e5ab97
|
3 |
-
size 3643
|
|
|
|
|
|
|
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2468874377
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d05009dbe23b1d997be5391df7f03167d76ffaee130c7dddc471568d4326068
|
3 |
size 2468874377
|