mHossain commited on
Commit
424bf40
Β·
1 Parent(s): 82fddf3

End of training

Browse files
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: csebuetnlp/mT5_m2m_crossSum
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - rouge
7
+ model-index:
8
+ - name: en_bn_summarize_v7
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # en_bn_summarize_v7
16
+
17
+ This model is a fine-tuned version of [csebuetnlp/mT5_m2m_crossSum](https://huggingface.co/csebuetnlp/mT5_m2m_crossSum) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: nan
20
+ - Rouge1: 0.0
21
+ - Rouge2: 0.0
22
+ - Rougel: 0.0
23
+ - Rougelsum: 0.0
24
+ - Gen Len: 28.882
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 2e-05
44
+ - train_batch_size: 4
45
+ - eval_batch_size: 4
46
+ - seed: 42
47
+ - gradient_accumulation_steps: 2
48
+ - total_train_batch_size: 8
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_steps: 5000
52
+ - num_epochs: 2
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
58
+ | No log | 1.0 | 154 | nan | 0.0 | 0.0 | 0.0 | 0.0 | 28.882 |
59
+ | No log | 2.0 | 308 | nan | 0.0 | 0.0 | 0.0 | 0.0 | 28.882 |
60
+
61
+
62
+ ### Framework versions
63
+
64
+ - Transformers 4.33.1
65
+ - Pytorch 2.0.1+cu118
66
+ - Datasets 2.14.5
67
+ - Tokenizers 0.13.3
config.json ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "csebuetnlp/mT5_m2m_crossSum",
3
+ "architectures": [
4
+ "MT5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 250030,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "length_penalty": 0.6,
20
+ "max_length": 84,
21
+ "model_type": "mt5",
22
+ "num_beams": 4,
23
+ "num_decoder_layers": 12,
24
+ "num_heads": 12,
25
+ "num_layers": 12,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "relative_attention_max_distance": 128,
29
+ "relative_attention_num_buckets": 32,
30
+ "task_specific_params": {
31
+ "langid_map": {
32
+ "amharic": [
33
+ 35,
34
+ "\u2581<extra_id_64>"
35
+ ],
36
+ "arabic": [
37
+ 4,
38
+ "\u2581<extra_id_95>"
39
+ ],
40
+ "azerbaijani": [
41
+ 7,
42
+ "\u2581<extra_id_92>"
43
+ ],
44
+ "bengali": [
45
+ 42,
46
+ "\u2581<extra_id_57>"
47
+ ],
48
+ "burmese": [
49
+ 33,
50
+ "\u2581<extra_id_66>"
51
+ ],
52
+ "chinese_simplified": [
53
+ 40,
54
+ "\u2581<extra_id_59>"
55
+ ],
56
+ "chinese_traditional": [
57
+ 44,
58
+ "\u2581<extra_id_55>"
59
+ ],
60
+ "english": [
61
+ 30,
62
+ "\u2581<extra_id_69>"
63
+ ],
64
+ "french": [
65
+ 10,
66
+ "\u2581<extra_id_89>"
67
+ ],
68
+ "gujarati": [
69
+ 27,
70
+ "\u2581<extra_id_72>"
71
+ ],
72
+ "hausa": [
73
+ 43,
74
+ "\u2581<extra_id_56>"
75
+ ],
76
+ "hindi": [
77
+ 21,
78
+ "\u2581<extra_id_78>"
79
+ ],
80
+ "igbo": [
81
+ 9,
82
+ "\u2581<extra_id_90>"
83
+ ],
84
+ "indonesian": [
85
+ 1,
86
+ "\u2581<extra_id_98>"
87
+ ],
88
+ "japanese": [
89
+ 37,
90
+ "\u2581<extra_id_62>"
91
+ ],
92
+ "kirundi": [
93
+ 0,
94
+ "\u2581<extra_id_99>"
95
+ ],
96
+ "korean": [
97
+ 29,
98
+ "\u2581<extra_id_70>"
99
+ ],
100
+ "kyrgyz": [
101
+ 5,
102
+ "\u2581<extra_id_94>"
103
+ ],
104
+ "marathi": [
105
+ 13,
106
+ "\u2581<extra_id_86>"
107
+ ],
108
+ "nepali": [
109
+ 20,
110
+ "\u2581<extra_id_79>"
111
+ ],
112
+ "oromo": [
113
+ 41,
114
+ "\u2581<extra_id_58>"
115
+ ],
116
+ "pashto": [
117
+ 34,
118
+ "\u2581<extra_id_65>"
119
+ ],
120
+ "persian": [
121
+ 23,
122
+ "\u2581<extra_id_76>"
123
+ ],
124
+ "pidgin": [
125
+ 14,
126
+ "\u2581<extra_id_85>"
127
+ ],
128
+ "portuguese": [
129
+ 39,
130
+ "\u2581<extra_id_60>"
131
+ ],
132
+ "punjabi": [
133
+ 17,
134
+ "\u2581<extra_id_82>"
135
+ ],
136
+ "russian": [
137
+ 36,
138
+ "\u2581<extra_id_63>"
139
+ ],
140
+ "scottish_gaelic": [
141
+ 24,
142
+ "\u2581<extra_id_75>"
143
+ ],
144
+ "serbian_cyrillic": [
145
+ 28,
146
+ "\u2581<extra_id_71>"
147
+ ],
148
+ "serbian_latin": [
149
+ 11,
150
+ "\u2581<extra_id_88>"
151
+ ],
152
+ "sinhala": [
153
+ 31,
154
+ "\u2581<extra_id_68>"
155
+ ],
156
+ "somali": [
157
+ 19,
158
+ "\u2581<extra_id_80>"
159
+ ],
160
+ "spanish": [
161
+ 3,
162
+ "\u2581<extra_id_96>"
163
+ ],
164
+ "swahili": [
165
+ 18,
166
+ "\u2581<extra_id_81>"
167
+ ],
168
+ "tamil": [
169
+ 32,
170
+ "\u2581<extra_id_67>"
171
+ ],
172
+ "telugu": [
173
+ 22,
174
+ "\u2581<extra_id_77>"
175
+ ],
176
+ "thai": [
177
+ 6,
178
+ "\u2581<extra_id_93>"
179
+ ],
180
+ "tigrinya": [
181
+ 16,
182
+ "\u2581<extra_id_83>"
183
+ ],
184
+ "turkish": [
185
+ 15,
186
+ "\u2581<extra_id_84>"
187
+ ],
188
+ "ukrainian": [
189
+ 2,
190
+ "\u2581<extra_id_97>"
191
+ ],
192
+ "urdu": [
193
+ 38,
194
+ "\u2581<extra_id_61>"
195
+ ],
196
+ "uzbek": [
197
+ 8,
198
+ "\u2581<extra_id_91>"
199
+ ],
200
+ "vietnamese": [
201
+ 12,
202
+ "\u2581<extra_id_87>"
203
+ ],
204
+ "welsh": [
205
+ 26,
206
+ "\u2581<extra_id_73>"
207
+ ],
208
+ "yoruba": [
209
+ 25,
210
+ "\u2581<extra_id_74>"
211
+ ]
212
+ }
213
+ },
214
+ "tie_word_embeddings": false,
215
+ "tokenizer_class": "T5Tokenizer",
216
+ "torch_dtype": "float32",
217
+ "transformers_version": "4.33.1",
218
+ "use_cache": true,
219
+ "vocab_size": 250112
220
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 250030,
3
+ "eos_token_id": 1,
4
+ "length_penalty": 0.6,
5
+ "max_length": 84,
6
+ "num_beams": 4,
7
+ "pad_token_id": 0,
8
+ "transformers_version": "4.33.1"
9
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b22fda21a0f5312002c077099ee460840f75ceb3842dd28483009c4984d96e1a
3
+ size 2329702581
special_tokens_map.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "▁<extra_id_64>",
4
+ "▁<extra_id_95>",
5
+ "▁<extra_id_92>",
6
+ "▁<extra_id_57>",
7
+ "▁<extra_id_66>",
8
+ "▁<extra_id_59>",
9
+ "▁<extra_id_55>",
10
+ "▁<extra_id_69>",
11
+ "▁<extra_id_89>",
12
+ "▁<extra_id_72>",
13
+ "▁<extra_id_56>",
14
+ "▁<extra_id_78>",
15
+ "▁<extra_id_90>",
16
+ "▁<extra_id_98>",
17
+ "▁<extra_id_62>",
18
+ "▁<extra_id_99>",
19
+ "▁<extra_id_70>",
20
+ "▁<extra_id_94>",
21
+ "▁<extra_id_86>",
22
+ "▁<extra_id_79>",
23
+ "▁<extra_id_58>",
24
+ "▁<extra_id_65>",
25
+ "▁<extra_id_76>",
26
+ "▁<extra_id_85>",
27
+ "▁<extra_id_60>",
28
+ "▁<extra_id_82>",
29
+ "▁<extra_id_63>",
30
+ "▁<extra_id_75>",
31
+ "▁<extra_id_71>",
32
+ "▁<extra_id_88>",
33
+ "▁<extra_id_68>",
34
+ "▁<extra_id_80>",
35
+ "▁<extra_id_96>",
36
+ "▁<extra_id_81>",
37
+ "▁<extra_id_67>",
38
+ "▁<extra_id_77>",
39
+ "▁<extra_id_93>",
40
+ "▁<extra_id_83>",
41
+ "▁<extra_id_84>",
42
+ "▁<extra_id_97>",
43
+ "▁<extra_id_61>",
44
+ "▁<extra_id_91>",
45
+ "▁<extra_id_87>",
46
+ "▁<extra_id_73>",
47
+ "▁<extra_id_74>"
48
+ ],
49
+ "eos_token": "</s>",
50
+ "pad_token": "<pad>",
51
+ "unk_token": "<unk>"
52
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": null,
3
+ "clean_up_tokenization_spaces": true,
4
+ "eos_token": "</s>",
5
+ "extra_ids": 0,
6
+ "legacy": true,
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<pad>",
9
+ "sp_model_kwargs": {},
10
+ "tokenizer_class": "T5Tokenizer",
11
+ "tokenizer_file": null,
12
+ "unk_token": "<unk>"
13
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2940ff1a2e61dd26c316012f9a5c7fa15f3424fea400d13a6a7f9b5a2b588837
3
+ size 4155