Muennighoff commited on
Commit
6e73714
·
1 Parent(s): 8ff6ce4

Add SGPT-1.3B-weightedmean-nli

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 2048,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": true,
8
+ "pooling_mode_lasttoken": false
9
+ }
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 2048 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `sentence_transformers.datasets.NoDuplicatesDataLoader.NoDuplicatesDataLoader` of length 93941 with parameters:
49
+ ```
50
+ {'batch_size': 6}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss` with parameters:
56
+ ```
57
+ {'scale': 20.0, 'similarity_fct': 'cos_sim'}
58
+ ```
59
+
60
+ Parameters of the fit()-Method:
61
+ ```
62
+ {
63
+ "epochs": 1,
64
+ "evaluation_steps": 9394,
65
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
66
+ "max_grad_norm": 1,
67
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
68
+ "optimizer_params": {
69
+ "lr": 1e-05
70
+ },
71
+ "scheduler": "WarmupLinear",
72
+ "steps_per_epoch": null,
73
+ "warmup_steps": 9395,
74
+ "weight_decay": 0.01
75
+ }
76
+ ```
77
+
78
+
79
+ ## Full Model Architecture
80
+ ```
81
+ SentenceTransformer(
82
+ (0): Transformer({'max_seq_length': 75, 'do_lower_case': False}) with Transformer model: GPTNeoModel
83
+ (1): Pooling({'word_embedding_dimension': 2048, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': True, 'pooling_mode_lasttoken': False})
84
+ )
85
+ ```
86
+
87
+ ## Citing & Authors
88
+
89
+ <!--- Describe where people can find more information -->
config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "EleutherAI/gpt-neo-1.3B",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPTNeoModel"
6
+ ],
7
+ "attention_dropout": 0,
8
+ "attention_layers": [
9
+ "global",
10
+ "local",
11
+ "global",
12
+ "local",
13
+ "global",
14
+ "local",
15
+ "global",
16
+ "local",
17
+ "global",
18
+ "local",
19
+ "global",
20
+ "local",
21
+ "global",
22
+ "local",
23
+ "global",
24
+ "local",
25
+ "global",
26
+ "local",
27
+ "global",
28
+ "local",
29
+ "global",
30
+ "local",
31
+ "global",
32
+ "local"
33
+ ],
34
+ "attention_types": [
35
+ [
36
+ [
37
+ "global",
38
+ "local"
39
+ ],
40
+ 12
41
+ ]
42
+ ],
43
+ "bos_token_id": 50256,
44
+ "embed_dropout": 0,
45
+ "eos_token_id": 50256,
46
+ "gradient_checkpointing": false,
47
+ "hidden_size": 2048,
48
+ "initializer_range": 0.02,
49
+ "intermediate_size": null,
50
+ "layer_norm_epsilon": 1e-05,
51
+ "max_position_embeddings": 2048,
52
+ "model_type": "gpt_neo",
53
+ "num_heads": 16,
54
+ "num_layers": 24,
55
+ "resid_dropout": 0,
56
+ "summary_activation": null,
57
+ "summary_first_dropout": 0.1,
58
+ "summary_proj_to_labels": true,
59
+ "summary_type": "cls_index",
60
+ "summary_use_proj": true,
61
+ "task_specific_params": {
62
+ "text-generation": {
63
+ "do_sample": true,
64
+ "max_length": 50,
65
+ "temperature": 0.9
66
+ }
67
+ },
68
+ "tokenizer_class": "GPT2Tokenizer",
69
+ "torch_dtype": "float32",
70
+ "transformers_version": "4.11.3",
71
+ "use_cache": true,
72
+ "vocab_size": 50257,
73
+ "window_size": 256
74
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.1.0",
4
+ "transformers": "4.11.3",
5
+ "pytorch": "1.10.1"
6
+ }
7
+ }
eval/similarity_evaluation_sts-dev_results.csv ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
3
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
4
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
5
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
6
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
7
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
8
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
9
+ 0,9394,0.8466169643976093,0.8572253068828245,0.8554214815746655,0.8581907470508644,0.8572737722326452,0.8609363902803336,0.7391976227377599,0.741672184433755
10
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
11
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
12
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
13
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
14
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
15
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
16
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
17
+ 0,18788,0.8469171819433461,0.8528195565114084,0.8305908574568376,0.8381609253988892,0.8306206110476602,0.8386161089870376,0.739936577843417,0.7527696582844111
18
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
19
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
20
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
21
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
22
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
23
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
24
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
25
+ 0,28182,0.8438525221445002,0.8488139683378565,0.8215216055658319,0.8294746578385671,0.8208760663530921,0.8300442311853723,0.7290087370710764,0.7481493657253322
26
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
27
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
28
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
29
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
30
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
31
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
32
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
33
+ 0,37576,0.8410092163771393,0.8434464254877504,0.8108116113116333,0.819470786166105,0.8114992962578211,0.8213302981255307,0.7233137320749693,0.7460423707236886
34
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
35
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
36
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
37
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
38
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
39
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
40
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
41
+ 0,46970,0.8379214058804858,0.840308129727673,0.8053616107375876,0.8133944616880712,0.8060678381432895,0.8146180781908651,0.7147880129476188,0.737420372546478
42
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
43
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
44
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
45
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
46
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
47
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
48
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
49
+ 0,56364,0.8391701725659262,0.843706124955228,0.8046337500463319,0.8134516313231266,0.804790182268625,0.8139750368218284,0.7189951349928259,0.7397492026796385
50
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
51
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
52
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
53
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
54
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
55
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
56
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
57
+ 0,65758,0.8344295349802213,0.8369230455410563,0.794231846207506,0.8046253559922896,0.7944726914504903,0.8051644318131093,0.7109246798299986,0.7341704728510499
58
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
59
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
60
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
61
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
62
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
63
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
64
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
65
+ 0,75152,0.8344864301248694,0.8370933125624741,0.7944898075152652,0.8041508307537827,0.7961352435570587,0.805845353275504,0.7104411626074354,0.7341434763911381
66
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
67
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
68
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
69
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
70
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
71
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
72
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
73
+ 0,84546,0.8308653030913641,0.8325898058958373,0.7866852395820515,0.7969515926894263,0.786763529391613,0.7969430873454784,0.7055077185906129,0.7310681361905212
74
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
75
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
76
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
77
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
78
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
79
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
80
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
81
+ 0,93940,0.8296429540019701,0.831143573581051,0.7841243612120149,0.7943857517671812,0.7845443517240553,0.794653970107371,0.7032306393600032,0.7305115628526874
82
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
83
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
84
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
85
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
86
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
87
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
88
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
89
+ 0,-1,0.8296504007134335,0.8311685647015004,0.7841439734651379,0.7944288667491779,0.7845598809192017,0.7947117032432987,0.7032215992698042,0.7304895663962262
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4bb338a5691f7536dc26acc4dfcd0023f563cc87c373703d6ab4d2800bdcede
3
+ size 5363096833
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 75,
3
+ "do_lower_case": false
4
+ }
similarity_evaluation_sts-test_results.csv ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
3
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
4
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
5
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
6
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
7
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
8
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
9
+ -1,-1,0.8215194634399462,0.831180286661736,0.8282469848503118,0.8260010179390973,0.830729410597307,0.8293667478450824,0.7055565453824795,0.6934185186991744
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>", "pad_token": "<|endoftext|>"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 2048, "special_tokens_map_file": null, "name_or_path": "EleutherAI/gpt-neo-1.3B", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff