Jyiyiyiyi commited on
Commit
cab0f89
·
verified ·
1 Parent(s): edfde5e

Upload 25 files

Browse files
0_Asym/1980970900352_Transformer/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../sts/output/training_ConMATH_ConMATH-2023-04-21_08-59-08\\0_Asym\\139753506388960_Transformer",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.41.2",
23
+ "vocab_size": 30527
24
+ }
0_Asym/1980970900352_Transformer/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:533fd7926e8f1c7c9571f6d0fb741fcefe20be5d09a394df306d6dc1c92970a5
3
+ size 437967672
0_Asym/1980970900352_Transformer/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
0_Asym/1980970900352_Transformer/special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
0_Asym/1980970900352_Transformer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
0_Asym/1980970900352_Transformer/tokenizer_config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[UNK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "30526": {
44
+ "content": "<mask>",
45
+ "lstrip": true,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ }
51
+ },
52
+ "bos_token": "<s>",
53
+ "clean_up_tokenization_spaces": true,
54
+ "cls_token": "<s>",
55
+ "do_lower_case": true,
56
+ "eos_token": "</s>",
57
+ "mask_token": "<mask>",
58
+ "max_length": 512,
59
+ "model_max_length": 512,
60
+ "pad_to_multiple_of": null,
61
+ "pad_token": "<pad>",
62
+ "pad_token_type_id": 0,
63
+ "padding_side": "right",
64
+ "sep_token": "</s>",
65
+ "stride": 0,
66
+ "strip_accents": null,
67
+ "tokenize_chinese_chars": true,
68
+ "tokenizer_class": "MPNetTokenizer",
69
+ "truncation_side": "right",
70
+ "truncation_strategy": "longest_first",
71
+ "unk_token": "[UNK]"
72
+ }
0_Asym/1980970900352_Transformer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
0_Asym/1980989322000_MarkuplmTransformerForConMATH/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<end-of-node>": 50266,
3
+ "[empty-title]": 50265
4
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../sts/output/training_ConMATH_ConMATH-2023-04-21_08-59-08\\0_Asym\\139751858333536_MarkuplmTransformerForConMATH",
3
+ "architectures": [
4
+ "MarkupLMModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_depth": 50,
18
+ "max_position_embeddings": 514,
19
+ "max_xpath_subs_unit_embeddings": 1024,
20
+ "max_xpath_tag_unit_embeddings": 256,
21
+ "model_type": "markuplm",
22
+ "num_attention_heads": 12,
23
+ "num_hidden_layers": 12,
24
+ "pad_token_id": 1,
25
+ "position_embedding_type": "absolute",
26
+ "subs_pad_id": 1001,
27
+ "tag_pad_id": 216,
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.41.2",
30
+ "type_vocab_size": 1,
31
+ "use_cache": true,
32
+ "vocab_size": 50267,
33
+ "xpath_unit_hidden_size": 32
34
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.41.2"
7
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
0_Asym/1980989322000_MarkuplmTransformerForConMATH/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a28d0e8d4efc67faf3400b21b061f74dd02ab4e37c24f2ed5b8a903314560cf2
3
+ size 540848992
0_Asym/1980989322000_MarkuplmTransformerForConMATH/preprocessor_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "feature_extractor_type": "MarkupLMFeatureExtractor",
3
+ "processor_class": "MarkupLMProcessor"
4
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
0_Asym/1980989322000_MarkuplmTransformerForConMATH/tokenizer_config.json ADDED
@@ -0,0 +1,283 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "50265": {
45
+ "content": "[empty-title]",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": false
51
+ },
52
+ "50266": {
53
+ "content": "<end-of-node>",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": false
59
+ }
60
+ },
61
+ "bos_token": "<s>",
62
+ "clean_up_tokenization_spaces": true,
63
+ "cls_token": "<s>",
64
+ "eos_token": "</s>",
65
+ "errors": "replace",
66
+ "mask_token": "<mask>",
67
+ "max_depth": 50,
68
+ "max_length": 512,
69
+ "max_width": 1000,
70
+ "model_max_length": 512,
71
+ "only_label_first_subword": true,
72
+ "pad_to_multiple_of": null,
73
+ "pad_token": "<pad>",
74
+ "pad_token_label": -100,
75
+ "pad_token_type_id": 0,
76
+ "pad_width": 1001,
77
+ "padding_side": "right",
78
+ "processor_class": "MarkupLMProcessor",
79
+ "sep_token": "</s>",
80
+ "stride": 0,
81
+ "tags_dict": {
82
+ "abs": 0,
83
+ "and": 1,
84
+ "annotation": 2,
85
+ "annotation-xml": 3,
86
+ "apply": 4,
87
+ "approx": 5,
88
+ "arccos": 6,
89
+ "arccosh": 7,
90
+ "arccot": 8,
91
+ "arccoth": 9,
92
+ "arccsc": 10,
93
+ "arccsch": 11,
94
+ "arcsec": 12,
95
+ "arcsech": 13,
96
+ "arcsin": 14,
97
+ "arcsinh": 15,
98
+ "arctan": 16,
99
+ "arctanh": 17,
100
+ "arg": 18,
101
+ "bind": 19,
102
+ "bvar": 20,
103
+ "card": 21,
104
+ "cartesianproduct": 22,
105
+ "cbytes": 23,
106
+ "ceiling": 24,
107
+ "cerror": 25,
108
+ "ci": 26,
109
+ "cn": 27,
110
+ "codomain": 28,
111
+ "complexes": 29,
112
+ "compose": 30,
113
+ "condition": 31,
114
+ "conjugate": 32,
115
+ "cos": 33,
116
+ "cosh": 34,
117
+ "cot": 35,
118
+ "coth": 36,
119
+ "cs": 37,
120
+ "csc": 38,
121
+ "csch": 39,
122
+ "csymbol": 40,
123
+ "curl": 41,
124
+ "declare": 42,
125
+ "degree": 43,
126
+ "determinant": 44,
127
+ "diff": 45,
128
+ "divergence": 46,
129
+ "divide": 47,
130
+ "domain": 48,
131
+ "domainofapplication": 49,
132
+ "emptyset": 50,
133
+ "eq": 51,
134
+ "equivalent": 52,
135
+ "eulergamma": 53,
136
+ "exists": 54,
137
+ "exp": 55,
138
+ "exponentiale": 56,
139
+ "factorial": 57,
140
+ "factorof": 58,
141
+ "false": 59,
142
+ "floor": 60,
143
+ "fn": 61,
144
+ "forall": 62,
145
+ "gcd": 63,
146
+ "geq": 64,
147
+ "grad": 65,
148
+ "gt": 66,
149
+ "ident": 67,
150
+ "image": 68,
151
+ "imaginary": 69,
152
+ "imaginaryi": 70,
153
+ "implies": 71,
154
+ "in": 72,
155
+ "infinity": 73,
156
+ "int": 74,
157
+ "integers": 75,
158
+ "intersect": 76,
159
+ "interval": 77,
160
+ "inverse": 78,
161
+ "lambda": 79,
162
+ "laplacian": 80,
163
+ "lcm": 81,
164
+ "leq": 82,
165
+ "limit": 83,
166
+ "list": 84,
167
+ "ln": 85,
168
+ "log": 86,
169
+ "logbase": 87,
170
+ "lowlimit": 88,
171
+ "lt": 89,
172
+ "maction": 90,
173
+ "maligngroup": 91,
174
+ "malignmark": 92,
175
+ "math": 93,
176
+ "matrix": 94,
177
+ "matrixrow": 95,
178
+ "max": 96,
179
+ "mean": 97,
180
+ "median": 98,
181
+ "menclose": 99,
182
+ "merror": 100,
183
+ "mfenced": 101,
184
+ "mfrac": 102,
185
+ "mglyph": 103,
186
+ "mi": 104,
187
+ "mi\"": 105,
188
+ "min": 106,
189
+ "minus": 107,
190
+ "mlabeledtr": 108,
191
+ "mlongdiv": 109,
192
+ "mmultiscripts": 110,
193
+ "mn": 111,
194
+ "mo": 112,
195
+ "mode": 113,
196
+ "moment": 114,
197
+ "momentabout": 115,
198
+ "mover": 116,
199
+ "mpadded": 117,
200
+ "mphantom": 118,
201
+ "mprescripts": 119,
202
+ "mroot": 120,
203
+ "mrow": 121,
204
+ "ms": 122,
205
+ "mscarries": 123,
206
+ "mscarry": 124,
207
+ "msgroup": 125,
208
+ "msline": 126,
209
+ "mspace": 127,
210
+ "msqrt": 128,
211
+ "msrow": 129,
212
+ "mstack": 130,
213
+ "mstyle": 131,
214
+ "msub": 132,
215
+ "msubsup": 133,
216
+ "msup": 134,
217
+ "mtable": 135,
218
+ "mtd": 136,
219
+ "mtext": 137,
220
+ "mtr": 138,
221
+ "munder": 139,
222
+ "munderover": 140,
223
+ "naturalnumbers": 141,
224
+ "neq": 142,
225
+ "none": 143,
226
+ "not": 144,
227
+ "notanumber": 145,
228
+ "notin": 146,
229
+ "notprsubset": 147,
230
+ "notsubset": 148,
231
+ "or": 149,
232
+ "otherwise": 150,
233
+ "outerproduct": 151,
234
+ "partialdiff": 152,
235
+ "pi": 153,
236
+ "piece": 154,
237
+ "piecewise": 155,
238
+ "plus": 156,
239
+ "power": 157,
240
+ "primes": 158,
241
+ "product": 159,
242
+ "prsubset": 160,
243
+ "quotient": 161,
244
+ "rationals": 162,
245
+ "real": 163,
246
+ "reals": 164,
247
+ "reln": 165,
248
+ "rem": 166,
249
+ "root": 167,
250
+ "scalarproduct": 168,
251
+ "sdev": 169,
252
+ "sec": 170,
253
+ "sech": 171,
254
+ "selector": 172,
255
+ "semantics": 173,
256
+ "sep": 174,
257
+ "set": 175,
258
+ "setdiff": 176,
259
+ "share": 177,
260
+ "sin": 178,
261
+ "sinh": 179,
262
+ "span": 180,
263
+ "subset": 181,
264
+ "sum": 182,
265
+ "tan": 183,
266
+ "tanh": 184,
267
+ "tendsto": 185,
268
+ "times": 186,
269
+ "transpose": 187,
270
+ "true": 188,
271
+ "union": 189,
272
+ "uplimit": 190,
273
+ "variance": 191,
274
+ "vector": 192,
275
+ "vectorproduct": 193,
276
+ "xor": 194
277
+ },
278
+ "tokenizer_class": "MarkupLMTokenizer",
279
+ "trim_offsets": false,
280
+ "truncation_side": "right",
281
+ "truncation_strategy": "longest_first",
282
+ "unk_token": "<unk>"
283
+ }
0_Asym/1980989322000_MarkuplmTransformerForConMATH/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
0_Asym/config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "types": {
3
+ "1980970900352_Transformer": "sentence_transformers.models.Transformer",
4
+ "1980989322000_MarkuplmTransformerForConMATH": "sentence_transformers.models.MarkuplmTransformerForConMATH"
5
+ },
6
+ "structure": {
7
+ "latex": [
8
+ "1980970900352_Transformer"
9
+ ],
10
+ "mathml": [
11
+ "1980989322000_MarkuplmTransformerForConMATH"
12
+ ]
13
+ },
14
+ "parameters": {
15
+ "allow_empty_key": true
16
+ }
17
+ }
1_Pooling/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false
9
+ }
2_Dense/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"in_features": 768, "out_features": 768, "bias": true, "activation_function": "torch.nn.modules.activation.Tanh"}
2_Dense/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b191669033cb2804deb580b26240d6821594240594a9d273b3ab37e3f05fa5b6
3
+ size 2363431
README.md ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+
8
+ ---
9
+
10
+ # {MODEL_NAME}
11
+
12
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
13
+
14
+ <!--- Describe your model here -->
15
+
16
+ ## Usage (Sentence-Transformers)
17
+
18
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
19
+
20
+ ```
21
+ pip install -U sentence-transformers
22
+ ```
23
+
24
+ Then you can use the model like this:
25
+
26
+ ```python
27
+ from sentence_transformers import SentenceTransformer
28
+ sentences = ["This is an example sentence", "Each sentence is converted"]
29
+
30
+ model = SentenceTransformer('{MODEL_NAME}')
31
+ embeddings = model.encode(sentences)
32
+ print(embeddings)
33
+ ```
34
+
35
+
36
+
37
+ ## Evaluation Results
38
+
39
+ <!--- Describe how your model was evaluated -->
40
+
41
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
42
+
43
+
44
+ ## Training
45
+ The model was trained with the parameters:
46
+
47
+ **DataLoader**:
48
+
49
+ `torch.utils.data.dataloader.DataLoader` of length 40 with parameters:
50
+ ```
51
+ {'batch_size': 4, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
52
+ ```
53
+
54
+ **Loss**:
55
+
56
+ `sentence_transformers.losses.MatryoshkaLoss.MatryoshkaLoss` with parameters:
57
+ ```
58
+ {'loss': 'MultipleNegativesSymmetricRankingLoss', 'matryoshka_dims': (768, 512, 256, 128, 64, 32, 16), 'matryoshka_weights': (1, 1, 1, 1, 1, 1, 1), 'n_dims_per_step': -1}
59
+ ```
60
+
61
+ Parameters of the fit()-Method:
62
+ ```
63
+ {
64
+ "epochs": 1,
65
+ "evaluation_steps": 4,
66
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
67
+ "max_grad_norm": 1,
68
+ "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
69
+ "optimizer_params": {
70
+ "lr": 5e-05
71
+ },
72
+ "scheduler": "WarmupLinear",
73
+ "steps_per_epoch": null,
74
+ "warmup_steps": 4,
75
+ "weight_decay": 0.01
76
+ }
77
+ ```
78
+
79
+
80
+ ## Full Model Architecture
81
+ ```
82
+ SentenceTransformer(
83
+ (0): Asym(
84
+ (latex-0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel
85
+ (mathml-0): MarkuplmTransformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MarkupLMModel
86
+ )
87
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False})
88
+ (2): Dense({'in_features': 768, 'out_features': 768, 'bias': True, 'activation_function': 'torch.nn.modules.activation.Tanh'})
89
+ )
90
+ ```
91
+
92
+ ## Citing & Authors
93
+
94
+ <!--- Describe where people can find more information -->
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.26.1",
5
+ "pytorch": "1.12.1+cu113"
6
+ }
7
+ }
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "0_Asym",
6
+ "type": "sentence_transformers.models.Asym"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Dense",
18
+ "type": "sentence_transformers.models.Dense"
19
+ }
20
+ ]