Add new SentenceTransformer model
Browse files- 1_Pooling/config.json +10 -0
- README.md +635 -0
- config.json +25 -0
- config_sentence_transformers.json +12 -0
- model.safetensors +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +63 -0
- vocab.txt +0 -0
1_Pooling/config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 1024,
|
3 |
+
"pooling_mode_cls_token": true,
|
4 |
+
"pooling_mode_mean_tokens": false,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
7 |
+
"pooling_mode_weightedmean_tokens": false,
|
8 |
+
"pooling_mode_lasttoken": false,
|
9 |
+
"include_prompt": true
|
10 |
+
}
|
README.md
ADDED
@@ -0,0 +1,635 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- sentence-transformers
|
4 |
+
- sentence-similarity
|
5 |
+
- feature-extraction
|
6 |
+
- generated_from_trainer
|
7 |
+
- dataset_size:410
|
8 |
+
- loss:MatryoshkaLoss
|
9 |
+
- loss:MultipleNegativesRankingLoss
|
10 |
+
base_model: Snowflake/snowflake-arctic-embed-l
|
11 |
+
widget:
|
12 |
+
- source_sentence: How did the LORD respond to Sarah's laughter and doubt about bearing
|
13 |
+
a child?
|
14 |
+
sentences:
|
15 |
+
- '"Stay here with the donkey; the boy and I will go over there; we will worship,
|
16 |
+
and then we will come back to you." [22:6] Abraham took the wood of the burnt
|
17 |
+
offering and laid it on his son Isaac, and he himself carried the fire and the
|
18 |
+
knife. So the two of them walked on together. [22:7] Isaac said to his father
|
19 |
+
Abraham, "Father!" And he said, "Here I am, my son." He said, "The fire and the
|
20 |
+
wood are here, but where is the lamb for a burnt offering?" [22:8] Abraham said,
|
21 |
+
"God himself will provide the lamb for a burnt offering, my son." So the two of
|
22 |
+
them walked on together. [22:9] When they came to the place that God had shown
|
23 |
+
him, Abraham built an altar there and laid the wood in order. He bound his son
|
24 |
+
Isaac, and laid him on the altar, on'
|
25 |
+
- you in due season, and your wife Sarah shall have a son." And Sarah was listening
|
26 |
+
at the tent entrance behind him. [18:11] Now Abraham and Sarah were old, advanced
|
27 |
+
in age; it had ceased to be with Sarah after the manner of women. [18:12] So Sarah
|
28 |
+
laughed to herself, saying, "After I have grown old, and my husband is old, shall
|
29 |
+
I have pleasure?" [18:13] The LORD said to Abraham, "Why did Sarah laugh, and
|
30 |
+
say, 'Shall I indeed bear a child, now that I am old?' [18:14] Is anything too
|
31 |
+
wonderful for the LORD? At the set time I will return to you, in due season, and
|
32 |
+
Sarah shall have a son." [18:15] But Sarah denied, saying, "I did not laugh";
|
33 |
+
for she was afraid. He said, "Oh yes, you did laugh." [18:16] Then the men set
|
34 |
+
out from there, and they
|
35 |
+
- face; perhaps he will accept me." [32:21] So the present passed on ahead of him;
|
36 |
+
and he himself spent that night in the camp. [32:22] The same night he got up
|
37 |
+
and took his two wives, his two maids, and his eleven children, and crossed the
|
38 |
+
ford of the Jabbok. [32:23] He took them and sent them across the stream, and
|
39 |
+
likewise everything that he had. [32:24] Jacob was left alone; and a man wrestled
|
40 |
+
with him until daybreak. [32:25] When the man saw that he did not prevail against
|
41 |
+
Jacob, he struck him on the hip socket; and Jacob's hip was put out of joint as
|
42 |
+
he wrestled with him. [32:26] Then he said, "Let me go, for the day is breaking."
|
43 |
+
But Jacob said, "I will not let you go, unless you bless me." [32:27] So he said
|
44 |
+
to him, "What is your
|
45 |
+
- source_sentence: What land does God promise to give to Abraham and his offspring?
|
46 |
+
sentences:
|
47 |
+
- for I have made you the ancestor of a multitude of nations. [17:6] I will make
|
48 |
+
you exceedingly fruitful; and I will make nations of you, and kings shall come
|
49 |
+
from you. [17:7] I will establish my covenant between me and you, and your offspring
|
50 |
+
after you throughout their generations, for an everlasting covenant, to be God
|
51 |
+
to you and to your offspring after you. [17:8] And I will give to you, and to
|
52 |
+
your offspring after you, the land where you are now an alien, all the land of
|
53 |
+
Canaan, for a perpetual holding; and I will be their God." [17:9] God said to
|
54 |
+
Abraham, "As for you, you shall keep my covenant, you and your offspring after
|
55 |
+
you throughout their generations. [17:10] This is my covenant, which you shall
|
56 |
+
keep, between me and you and your
|
57 |
+
- and his mother prepared savory food, such as his father loved. [27:15] Then Rebekah
|
58 |
+
took the best garments of her elder son Esau, which were with her in the house,
|
59 |
+
and put them on her younger son Jacob; [27:16] and she put the skins of the kids
|
60 |
+
on his hands and on the smooth part of his neck. [27:17] Then she handed the savory
|
61 |
+
food, and the bread that she had prepared, to her son Jacob. [27:18] So he went
|
62 |
+
in to his father, and said, "My father"; and he said, "Here I am; who are you,
|
63 |
+
my son?" [27:19] Jacob said to his father, "I am Esau your firstborn. I have done
|
64 |
+
as you told me; now sit up and eat of my game, so that you may bless me." [27:20]
|
65 |
+
But Isaac said to his son, "How is it that you have found it so quickly, my son?"
|
66 |
+
He answered,
|
67 |
+
- you for a burying place, so that I may bury my dead out of my sight." [23:5] The
|
68 |
+
Hittites answered Abraham, [23:6] "Hear us, my lord; you are a mighty prince among
|
69 |
+
us. Bury your dead in the choicest of our burial places; none of us will withhold
|
70 |
+
from you any burial ground for burying your dead." [23:7] Abraham rose and bowed
|
71 |
+
to the Hittites, the people of the land. [23:8] He said to them, "If you are willing
|
72 |
+
that I should bury my dead out of my sight, hear me, and entreat for me Ephron
|
73 |
+
son of Zohar, [23:9] so that he may give me the cave of Machpelah, which he owns;
|
74 |
+
it is at the end of his field. For the full price let him give it to me in your
|
75 |
+
presence as a possession for a burying place." [23:10] Now Ephron was sitting
|
76 |
+
among the
|
77 |
+
- source_sentence: At what age did Enosh become the father of Kenan?
|
78 |
+
sentences:
|
79 |
+
- of Egypt to the great river, the river Euphrates, [15:19] the land of the Kenites,
|
80 |
+
the Kenizzites, the Kadmonites, [15:20] the Hittites, the Perizzites, the Rephaim,
|
81 |
+
[15:21] the Amorites, the Canaanites, the Girgashites, and the Jebusites.". Chapter
|
82 |
+
16 [16:1] Now Sarai, Abram's wife, bore him no children. She had an Egyptian slave-girl
|
83 |
+
whose name was Hagar, [16:2] and Sarai said to Abram, "You see that the LORD has
|
84 |
+
prevented me from bearing children; go in to my slave-girl; it may be that I shall
|
85 |
+
obtain children by her." And Abram listened to the voice of Sarai. [16:3] So,
|
86 |
+
after Abram had lived ten years in the land of Canaan, Sarai, Abram's wife, took
|
87 |
+
Hagar the Egyptian, her slave-girl, and gave her to her husband Abram as a wife.
|
88 |
+
[16:4]
|
89 |
+
- to his image, and named him Seth. [5:4] The days of Adam after he became the father
|
90 |
+
of Seth were eight hundred years; and he had other sons and daughters. [5:5] Thus
|
91 |
+
all the days that Adam lived were nine hundred thirty years; and he died. [5:6]
|
92 |
+
When Seth had lived one hundred five years, he became the father of Enosh. [5:7]
|
93 |
+
Seth lived after the birth of Enosh eight hundred seven years, and had other sons
|
94 |
+
and daughters. [5:8] Thus all the days of Seth were nine hundred twelve years;
|
95 |
+
and he died. [5:9] When Enosh had lived ninety years, he became the father of
|
96 |
+
Kenan. [5:10] Enosh lived after the birth of Kenan eight hundred fifteen years,
|
97 |
+
and had other sons and daughters. [5:11] Thus all the days of Enosh were nine
|
98 |
+
hundred five years; and
|
99 |
+
- said, "Come, let us build ourselves a city, and a tower with its top in the heavens,
|
100 |
+
and let us make a name for ourselves; otherwise we shall be scattered abroad upon
|
101 |
+
the face of the whole earth." [11:5] The LORD came down to see the city and the
|
102 |
+
tower, which mortals had built. [11:6] And the LORD said, "Look, they are one
|
103 |
+
people, and they have all one language; and this is only the beginning of what
|
104 |
+
they will do; nothing that they propose to do will now be impossible for them.
|
105 |
+
[11:7] Come, let us go down, and confuse their language there, so that they will
|
106 |
+
not understand one another's speech." [11:8] So the LORD scattered them abroad
|
107 |
+
from there over the face of all the earth, and they left off building the city.
|
108 |
+
[11:9] Therefore it was
|
109 |
+
- source_sentence: How did the angels assist Lot and his family in escaping the city?
|
110 |
+
sentences:
|
111 |
+
- has become great before the LORD, and the LORD has sent us to destroy it." [19:14]
|
112 |
+
So Lot went out and said to his sons-in-law, who were to marry his daughters,
|
113 |
+
"Up, get out of this place; for the LORD is about to destroy the city." But he
|
114 |
+
seemed to his sons-in-law to be jesting. [19:15] When morning dawned, the angels
|
115 |
+
urged Lot, saying, "Get up, take your wife and your two daughters who are here,
|
116 |
+
or else you will be consumed in the punishment of the city." [19:16] But he lingered;
|
117 |
+
so the men seized him and his wife and his two daughters by the hand, the LORD
|
118 |
+
being merciful to him, and they brought him out and left him outside the city.
|
119 |
+
[19:17] When they had brought them outside, they said, "Flee for your life; do
|
120 |
+
not look back or stop
|
121 |
+
- five years; and he died. [5:12] When Kenan had lived seventy years, he became
|
122 |
+
the father of Mahalalel. [5:13] Kenan lived after the birth of Mahalalel eight
|
123 |
+
hundred and forty years, and had other sons and daughters. [5:14] Thus all the
|
124 |
+
days of Kenan were nine hundred and ten years; and he died. [5:15] When Mahalalel
|
125 |
+
had lived sixty-five years, he became the father of Jared. [5:16] Mahalalel lived
|
126 |
+
after the birth of Jared eight hundred thirty years, and had other sons and daughters.
|
127 |
+
[5:17] Thus all the days of Mahalalel were eight hundred ninety-five years; and
|
128 |
+
he died. [5:18] When Jared had lived one hundred sixty-two years he became the
|
129 |
+
father of Enoch. [5:19] Jared lived after the birth of Enoch eight hundred years,
|
130 |
+
and had other sons
|
131 |
+
- go with this man?" She said, "I will." [24:59] So they sent away their sister
|
132 |
+
Rebekah and her nurse along with Abraham's servant and his men. [24:60] And they
|
133 |
+
blessed Rebekah and said to her, "May you, our sister, become thousands of myriads;
|
134 |
+
may your offspring gain possession of the gates of their foes." [24:61] Then Rebekah
|
135 |
+
and her maids rose up, mounted the camels, and followed the man; thus the servant
|
136 |
+
took Rebekah, and went his way. [24:62] Now Isaac had come from Beer-lahai-roi,
|
137 |
+
and was settled in the Negeb. [24:63] Isaac went out in the evening to walk in
|
138 |
+
the field; and looking up, he saw camels coming. [24:64] And Rebekah looked up,
|
139 |
+
and when she saw Isaac, she slipped quickly from the camel, [24:65] and said to
|
140 |
+
the servant, "Who is
|
141 |
+
- source_sentence: What did Abraham serve to the visitors while they ate under the
|
142 |
+
tree?
|
143 |
+
sentences:
|
144 |
+
- '[21:34] And Abraham resided as an alien many days in the land of the Philistines. Chapter
|
145 |
+
22 [22:1] After these things God tested Abraham. He said to him, "Abraham!" And
|
146 |
+
he said, "Here I am." [22:2] He said, "Take your son, your only son Isaac, whom
|
147 |
+
you love, and go to the land of Moriah, and offer him there as a burnt offering
|
148 |
+
on one of the mountains that I shall show you." [22:3] So Abraham rose early in
|
149 |
+
the morning, saddled his donkey, and took two of his young men with him, and his
|
150 |
+
son Isaac; he cut the wood for the burnt offering, and set out and went to the
|
151 |
+
place in the distance that God had shown him. [22:4] On the third day Abraham
|
152 |
+
looked up and saw the place far away. [22:5] Then Abraham said to his young men,
|
153 |
+
"Stay here with the'
|
154 |
+
- tree. [18:5] Let me bring a little bread, that you may refresh yourselves, and
|
155 |
+
after that you may pass on - since you have come to your servant." So they said,
|
156 |
+
"Do as you have said." [18:6] And Abraham hastened into the tent to Sarah, and
|
157 |
+
said, "Make ready quickly three measures of choice flour, knead it, and make cakes.
|
158 |
+
" [18:7] Abraham ran to the herd, and took a calf, tender and good, and gave it
|
159 |
+
to the servant, who hastened to prepare it. [18:8] Then he took curds and milk
|
160 |
+
and the calf that he had prepared, and set it before them; and he stood by them
|
161 |
+
under the tree while they ate. [18:9] They said to him, "Where is your wife Sarah?"
|
162 |
+
And he said, "There, in the tent." [18:10] Then one said, "I will surely return
|
163 |
+
to you in due season,
|
164 |
+
- '[30:24] and she named him Joseph, saying, "May the LORD add to me another son!"
|
165 |
+
[30:25] When Rachel had borne Joseph, Jacob said to Laban, "Send me away, that
|
166 |
+
I may go to my own home and country. [30:26] Give me my wives and my children
|
167 |
+
for whom I have served you, and let me go; for you know very well the service
|
168 |
+
I have given you." [30:27] But Laban said to him, "If you will allow me to say
|
169 |
+
so, I have learned by divination that the LORD has blessed me because of you;
|
170 |
+
[30:28] name your wages, and I will give it." [30:29] Jacob said to him, "You
|
171 |
+
yourself know how I have served you, and how your cattle have fared with me. [30:30]
|
172 |
+
For you had little before I came, and it has increased abundantly; and the LORD
|
173 |
+
has blessed you wherever I turned.'
|
174 |
+
pipeline_tag: sentence-similarity
|
175 |
+
library_name: sentence-transformers
|
176 |
+
metrics:
|
177 |
+
- cosine_accuracy@1
|
178 |
+
- cosine_accuracy@3
|
179 |
+
- cosine_accuracy@5
|
180 |
+
- cosine_accuracy@10
|
181 |
+
- cosine_precision@1
|
182 |
+
- cosine_precision@3
|
183 |
+
- cosine_precision@5
|
184 |
+
- cosine_precision@10
|
185 |
+
- cosine_recall@1
|
186 |
+
- cosine_recall@3
|
187 |
+
- cosine_recall@5
|
188 |
+
- cosine_recall@10
|
189 |
+
- cosine_ndcg@10
|
190 |
+
- cosine_mrr@10
|
191 |
+
- cosine_map@100
|
192 |
+
model-index:
|
193 |
+
- name: SentenceTransformer based on Snowflake/snowflake-arctic-embed-l
|
194 |
+
results:
|
195 |
+
- task:
|
196 |
+
type: information-retrieval
|
197 |
+
name: Information Retrieval
|
198 |
+
dataset:
|
199 |
+
name: Unknown
|
200 |
+
type: unknown
|
201 |
+
metrics:
|
202 |
+
- type: cosine_accuracy@1
|
203 |
+
value: 0.75
|
204 |
+
name: Cosine Accuracy@1
|
205 |
+
- type: cosine_accuracy@3
|
206 |
+
value: 0.9375
|
207 |
+
name: Cosine Accuracy@3
|
208 |
+
- type: cosine_accuracy@5
|
209 |
+
value: 0.975
|
210 |
+
name: Cosine Accuracy@5
|
211 |
+
- type: cosine_accuracy@10
|
212 |
+
value: 0.9875
|
213 |
+
name: Cosine Accuracy@10
|
214 |
+
- type: cosine_precision@1
|
215 |
+
value: 0.75
|
216 |
+
name: Cosine Precision@1
|
217 |
+
- type: cosine_precision@3
|
218 |
+
value: 0.3125
|
219 |
+
name: Cosine Precision@3
|
220 |
+
- type: cosine_precision@5
|
221 |
+
value: 0.19499999999999998
|
222 |
+
name: Cosine Precision@5
|
223 |
+
- type: cosine_precision@10
|
224 |
+
value: 0.09874999999999998
|
225 |
+
name: Cosine Precision@10
|
226 |
+
- type: cosine_recall@1
|
227 |
+
value: 0.75
|
228 |
+
name: Cosine Recall@1
|
229 |
+
- type: cosine_recall@3
|
230 |
+
value: 0.9375
|
231 |
+
name: Cosine Recall@3
|
232 |
+
- type: cosine_recall@5
|
233 |
+
value: 0.975
|
234 |
+
name: Cosine Recall@5
|
235 |
+
- type: cosine_recall@10
|
236 |
+
value: 0.9875
|
237 |
+
name: Cosine Recall@10
|
238 |
+
- type: cosine_ndcg@10
|
239 |
+
value: 0.8820698787104944
|
240 |
+
name: Cosine Ndcg@10
|
241 |
+
- type: cosine_mrr@10
|
242 |
+
value: 0.8465773809523809
|
243 |
+
name: Cosine Mrr@10
|
244 |
+
- type: cosine_map@100
|
245 |
+
value: 0.8472718253968254
|
246 |
+
name: Cosine Map@100
|
247 |
+
---
|
248 |
+
|
249 |
+
# SentenceTransformer based on Snowflake/snowflake-arctic-embed-l
|
250 |
+
|
251 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Snowflake/snowflake-arctic-embed-l](https://huggingface.co/Snowflake/snowflake-arctic-embed-l). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
252 |
+
|
253 |
+
## Model Details
|
254 |
+
|
255 |
+
### Model Description
|
256 |
+
- **Model Type:** Sentence Transformer
|
257 |
+
- **Base model:** [Snowflake/snowflake-arctic-embed-l](https://huggingface.co/Snowflake/snowflake-arctic-embed-l) <!-- at revision d8fb21ca8d905d2832ee8b96c894d3298964346b -->
|
258 |
+
- **Maximum Sequence Length:** 512 tokens
|
259 |
+
- **Output Dimensionality:** 1024 dimensions
|
260 |
+
- **Similarity Function:** Cosine Similarity
|
261 |
+
<!-- - **Training Dataset:** Unknown -->
|
262 |
+
<!-- - **Language:** Unknown -->
|
263 |
+
<!-- - **License:** Unknown -->
|
264 |
+
|
265 |
+
### Model Sources
|
266 |
+
|
267 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
268 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
269 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
270 |
+
|
271 |
+
### Full Model Architecture
|
272 |
+
|
273 |
+
```
|
274 |
+
SentenceTransformer(
|
275 |
+
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
|
276 |
+
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
277 |
+
(2): Normalize()
|
278 |
+
)
|
279 |
+
```
|
280 |
+
|
281 |
+
## Usage
|
282 |
+
|
283 |
+
### Direct Usage (Sentence Transformers)
|
284 |
+
|
285 |
+
First install the Sentence Transformers library:
|
286 |
+
|
287 |
+
```bash
|
288 |
+
pip install -U sentence-transformers
|
289 |
+
```
|
290 |
+
|
291 |
+
Then you can load this model and run inference.
|
292 |
+
```python
|
293 |
+
from sentence_transformers import SentenceTransformer
|
294 |
+
|
295 |
+
# Download from the 🤗 Hub
|
296 |
+
model = SentenceTransformer("kcheng0816/finetuned_arctic_genesis")
|
297 |
+
# Run inference
|
298 |
+
sentences = [
|
299 |
+
'What did Abraham serve to the visitors while they ate under the tree?',
|
300 |
+
'tree. [18:5] Let me bring a little bread, that you may refresh yourselves, and after that you may pass on - since you have come to your servant." So they said, "Do as you have said." [18:6] And Abraham hastened into the tent to Sarah, and said, "Make ready quickly three measures of choice flour, knead it, and make cakes. " [18:7] Abraham ran to the herd, and took a calf, tender and good, and gave it to the servant, who hastened to prepare it. [18:8] Then he took curds and milk and the calf that he had prepared, and set it before them; and he stood by them under the tree while they ate. [18:9] They said to him, "Where is your wife Sarah?" And he said, "There, in the tent." [18:10] Then one said, "I will surely return to you in due season,',
|
301 |
+
'[21:34] And Abraham resided as an alien many days in the land of the Philistines. Chapter 22 [22:1] After these things God tested Abraham. He said to him, "Abraham!" And he said, "Here I am." [22:2] He said, "Take your son, your only son Isaac, whom you love, and go to the land of Moriah, and offer him there as a burnt offering on one of the mountains that I shall show you." [22:3] So Abraham rose early in the morning, saddled his donkey, and took two of his young men with him, and his son Isaac; he cut the wood for the burnt offering, and set out and went to the place in the distance that God had shown him. [22:4] On the third day Abraham looked up and saw the place far away. [22:5] Then Abraham said to his young men, "Stay here with the',
|
302 |
+
]
|
303 |
+
embeddings = model.encode(sentences)
|
304 |
+
print(embeddings.shape)
|
305 |
+
# [3, 1024]
|
306 |
+
|
307 |
+
# Get the similarity scores for the embeddings
|
308 |
+
similarities = model.similarity(embeddings, embeddings)
|
309 |
+
print(similarities.shape)
|
310 |
+
# [3, 3]
|
311 |
+
```
|
312 |
+
|
313 |
+
<!--
|
314 |
+
### Direct Usage (Transformers)
|
315 |
+
|
316 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
317 |
+
|
318 |
+
</details>
|
319 |
+
-->
|
320 |
+
|
321 |
+
<!--
|
322 |
+
### Downstream Usage (Sentence Transformers)
|
323 |
+
|
324 |
+
You can finetune this model on your own dataset.
|
325 |
+
|
326 |
+
<details><summary>Click to expand</summary>
|
327 |
+
|
328 |
+
</details>
|
329 |
+
-->
|
330 |
+
|
331 |
+
<!--
|
332 |
+
### Out-of-Scope Use
|
333 |
+
|
334 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
335 |
+
-->
|
336 |
+
|
337 |
+
## Evaluation
|
338 |
+
|
339 |
+
### Metrics
|
340 |
+
|
341 |
+
#### Information Retrieval
|
342 |
+
|
343 |
+
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
344 |
+
|
345 |
+
| Metric | Value |
|
346 |
+
|:--------------------|:-----------|
|
347 |
+
| cosine_accuracy@1 | 0.75 |
|
348 |
+
| cosine_accuracy@3 | 0.9375 |
|
349 |
+
| cosine_accuracy@5 | 0.975 |
|
350 |
+
| cosine_accuracy@10 | 0.9875 |
|
351 |
+
| cosine_precision@1 | 0.75 |
|
352 |
+
| cosine_precision@3 | 0.3125 |
|
353 |
+
| cosine_precision@5 | 0.195 |
|
354 |
+
| cosine_precision@10 | 0.0987 |
|
355 |
+
| cosine_recall@1 | 0.75 |
|
356 |
+
| cosine_recall@3 | 0.9375 |
|
357 |
+
| cosine_recall@5 | 0.975 |
|
358 |
+
| cosine_recall@10 | 0.9875 |
|
359 |
+
| **cosine_ndcg@10** | **0.8821** |
|
360 |
+
| cosine_mrr@10 | 0.8466 |
|
361 |
+
| cosine_map@100 | 0.8473 |
|
362 |
+
|
363 |
+
<!--
|
364 |
+
## Bias, Risks and Limitations
|
365 |
+
|
366 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
367 |
+
-->
|
368 |
+
|
369 |
+
<!--
|
370 |
+
### Recommendations
|
371 |
+
|
372 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
373 |
+
-->
|
374 |
+
|
375 |
+
## Training Details
|
376 |
+
|
377 |
+
### Training Dataset
|
378 |
+
|
379 |
+
#### Unnamed Dataset
|
380 |
+
|
381 |
+
* Size: 410 training samples
|
382 |
+
* Columns: <code>sentence_0</code> and <code>sentence_1</code>
|
383 |
+
* Approximate statistics based on the first 410 samples:
|
384 |
+
| | sentence_0 | sentence_1 |
|
385 |
+
|:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
|
386 |
+
| type | string | string |
|
387 |
+
| details | <ul><li>min: 10 tokens</li><li>mean: 17.63 tokens</li><li>max: 31 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 206.17 tokens</li><li>max: 257 tokens</li></ul> |
|
388 |
+
* Samples:
|
389 |
+
| sentence_0 | sentence_1 |
|
390 |
+
|:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
391 |
+
| <code>What are the main themes explored in the Book of Genesis?</code> | <code>The Book of Genesis</code> |
|
392 |
+
| <code>How does the Book of Genesis describe the creation of the world?</code> | <code>The Book of Genesis</code> |
|
393 |
+
| <code>What did God create in the beginning according to the Book of Genesis?</code> | <code>THE BOOK OF GENESIS 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50 Chapter 1 [1:1] In the beginning when God created the heavens and the earth, [1:2] the earth was a formless void and darkness covered the face of the deep, while a wind from God swept over the face of the waters. [1:3] Then God said, "Let there be light"; and there was light. [1:4] And God saw that the light was good; and God separated the light from the darkness. [1:5] God called the light Day, and the darkness he called Night. And there was evening and there was morning, the first day. [1:6] And God said, "Let there be</code> |
|
394 |
+
* Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
|
395 |
+
```json
|
396 |
+
{
|
397 |
+
"loss": "MultipleNegativesRankingLoss",
|
398 |
+
"matryoshka_dims": [
|
399 |
+
768,
|
400 |
+
512,
|
401 |
+
256,
|
402 |
+
128,
|
403 |
+
64
|
404 |
+
],
|
405 |
+
"matryoshka_weights": [
|
406 |
+
1,
|
407 |
+
1,
|
408 |
+
1,
|
409 |
+
1,
|
410 |
+
1
|
411 |
+
],
|
412 |
+
"n_dims_per_step": -1
|
413 |
+
}
|
414 |
+
```
|
415 |
+
|
416 |
+
### Training Hyperparameters
|
417 |
+
#### Non-Default Hyperparameters
|
418 |
+
|
419 |
+
- `eval_strategy`: steps
|
420 |
+
- `per_device_train_batch_size`: 10
|
421 |
+
- `per_device_eval_batch_size`: 10
|
422 |
+
- `num_train_epochs`: 10
|
423 |
+
- `multi_dataset_batch_sampler`: round_robin
|
424 |
+
|
425 |
+
#### All Hyperparameters
|
426 |
+
<details><summary>Click to expand</summary>
|
427 |
+
|
428 |
+
- `overwrite_output_dir`: False
|
429 |
+
- `do_predict`: False
|
430 |
+
- `eval_strategy`: steps
|
431 |
+
- `prediction_loss_only`: True
|
432 |
+
- `per_device_train_batch_size`: 10
|
433 |
+
- `per_device_eval_batch_size`: 10
|
434 |
+
- `per_gpu_train_batch_size`: None
|
435 |
+
- `per_gpu_eval_batch_size`: None
|
436 |
+
- `gradient_accumulation_steps`: 1
|
437 |
+
- `eval_accumulation_steps`: None
|
438 |
+
- `torch_empty_cache_steps`: None
|
439 |
+
- `learning_rate`: 5e-05
|
440 |
+
- `weight_decay`: 0.0
|
441 |
+
- `adam_beta1`: 0.9
|
442 |
+
- `adam_beta2`: 0.999
|
443 |
+
- `adam_epsilon`: 1e-08
|
444 |
+
- `max_grad_norm`: 1
|
445 |
+
- `num_train_epochs`: 10
|
446 |
+
- `max_steps`: -1
|
447 |
+
- `lr_scheduler_type`: linear
|
448 |
+
- `lr_scheduler_kwargs`: {}
|
449 |
+
- `warmup_ratio`: 0.0
|
450 |
+
- `warmup_steps`: 0
|
451 |
+
- `log_level`: passive
|
452 |
+
- `log_level_replica`: warning
|
453 |
+
- `log_on_each_node`: True
|
454 |
+
- `logging_nan_inf_filter`: True
|
455 |
+
- `save_safetensors`: True
|
456 |
+
- `save_on_each_node`: False
|
457 |
+
- `save_only_model`: False
|
458 |
+
- `restore_callback_states_from_checkpoint`: False
|
459 |
+
- `no_cuda`: False
|
460 |
+
- `use_cpu`: False
|
461 |
+
- `use_mps_device`: False
|
462 |
+
- `seed`: 42
|
463 |
+
- `data_seed`: None
|
464 |
+
- `jit_mode_eval`: False
|
465 |
+
- `use_ipex`: False
|
466 |
+
- `bf16`: False
|
467 |
+
- `fp16`: False
|
468 |
+
- `fp16_opt_level`: O1
|
469 |
+
- `half_precision_backend`: auto
|
470 |
+
- `bf16_full_eval`: False
|
471 |
+
- `fp16_full_eval`: False
|
472 |
+
- `tf32`: None
|
473 |
+
- `local_rank`: 0
|
474 |
+
- `ddp_backend`: None
|
475 |
+
- `tpu_num_cores`: None
|
476 |
+
- `tpu_metrics_debug`: False
|
477 |
+
- `debug`: []
|
478 |
+
- `dataloader_drop_last`: False
|
479 |
+
- `dataloader_num_workers`: 0
|
480 |
+
- `dataloader_prefetch_factor`: None
|
481 |
+
- `past_index`: -1
|
482 |
+
- `disable_tqdm`: False
|
483 |
+
- `remove_unused_columns`: True
|
484 |
+
- `label_names`: None
|
485 |
+
- `load_best_model_at_end`: False
|
486 |
+
- `ignore_data_skip`: False
|
487 |
+
- `fsdp`: []
|
488 |
+
- `fsdp_min_num_params`: 0
|
489 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
490 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
491 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
492 |
+
- `deepspeed`: None
|
493 |
+
- `label_smoothing_factor`: 0.0
|
494 |
+
- `optim`: adamw_torch
|
495 |
+
- `optim_args`: None
|
496 |
+
- `adafactor`: False
|
497 |
+
- `group_by_length`: False
|
498 |
+
- `length_column_name`: length
|
499 |
+
- `ddp_find_unused_parameters`: None
|
500 |
+
- `ddp_bucket_cap_mb`: None
|
501 |
+
- `ddp_broadcast_buffers`: False
|
502 |
+
- `dataloader_pin_memory`: True
|
503 |
+
- `dataloader_persistent_workers`: False
|
504 |
+
- `skip_memory_metrics`: True
|
505 |
+
- `use_legacy_prediction_loop`: False
|
506 |
+
- `push_to_hub`: False
|
507 |
+
- `resume_from_checkpoint`: None
|
508 |
+
- `hub_model_id`: None
|
509 |
+
- `hub_strategy`: every_save
|
510 |
+
- `hub_private_repo`: None
|
511 |
+
- `hub_always_push`: False
|
512 |
+
- `gradient_checkpointing`: False
|
513 |
+
- `gradient_checkpointing_kwargs`: None
|
514 |
+
- `include_inputs_for_metrics`: False
|
515 |
+
- `include_for_metrics`: []
|
516 |
+
- `eval_do_concat_batches`: True
|
517 |
+
- `fp16_backend`: auto
|
518 |
+
- `push_to_hub_model_id`: None
|
519 |
+
- `push_to_hub_organization`: None
|
520 |
+
- `mp_parameters`:
|
521 |
+
- `auto_find_batch_size`: False
|
522 |
+
- `full_determinism`: False
|
523 |
+
- `torchdynamo`: None
|
524 |
+
- `ray_scope`: last
|
525 |
+
- `ddp_timeout`: 1800
|
526 |
+
- `torch_compile`: False
|
527 |
+
- `torch_compile_backend`: None
|
528 |
+
- `torch_compile_mode`: None
|
529 |
+
- `dispatch_batches`: None
|
530 |
+
- `split_batches`: None
|
531 |
+
- `include_tokens_per_second`: False
|
532 |
+
- `include_num_input_tokens_seen`: False
|
533 |
+
- `neftune_noise_alpha`: None
|
534 |
+
- `optim_target_modules`: None
|
535 |
+
- `batch_eval_metrics`: False
|
536 |
+
- `eval_on_start`: False
|
537 |
+
- `use_liger_kernel`: False
|
538 |
+
- `eval_use_gather_object`: False
|
539 |
+
- `average_tokens_across_devices`: False
|
540 |
+
- `prompts`: None
|
541 |
+
- `batch_sampler`: batch_sampler
|
542 |
+
- `multi_dataset_batch_sampler`: round_robin
|
543 |
+
|
544 |
+
</details>
|
545 |
+
|
546 |
+
### Training Logs
|
547 |
+
| Epoch | Step | cosine_ndcg@10 |
|
548 |
+
|:------:|:----:|:--------------:|
|
549 |
+
| 1.0 | 41 | 0.8988 |
|
550 |
+
| 1.2195 | 50 | 0.8824 |
|
551 |
+
| 2.0 | 82 | 0.8775 |
|
552 |
+
| 2.4390 | 100 | 0.8808 |
|
553 |
+
| 3.0 | 123 | 0.8673 |
|
554 |
+
| 3.6585 | 150 | 0.8634 |
|
555 |
+
| 4.0 | 164 | 0.8735 |
|
556 |
+
| 4.8780 | 200 | 0.8730 |
|
557 |
+
| 5.0 | 205 | 0.8713 |
|
558 |
+
| 6.0 | 246 | 0.8719 |
|
559 |
+
| 6.0976 | 250 | 0.8765 |
|
560 |
+
| 7.0 | 287 | 0.8848 |
|
561 |
+
| 7.3171 | 300 | 0.8783 |
|
562 |
+
| 8.0 | 328 | 0.8892 |
|
563 |
+
| 8.5366 | 350 | 0.8881 |
|
564 |
+
| 9.0 | 369 | 0.8821 |
|
565 |
+
| 9.7561 | 400 | 0.8821 |
|
566 |
+
| 10.0 | 410 | 0.8821 |
|
567 |
+
|
568 |
+
|
569 |
+
### Framework Versions
|
570 |
+
- Python: 3.11.11
|
571 |
+
- Sentence Transformers: 3.4.1
|
572 |
+
- Transformers: 4.49.0
|
573 |
+
- PyTorch: 2.6.0
|
574 |
+
- Accelerate: 1.3.0
|
575 |
+
- Datasets: 3.3.2
|
576 |
+
- Tokenizers: 0.21.0
|
577 |
+
|
578 |
+
## Citation
|
579 |
+
|
580 |
+
### BibTeX
|
581 |
+
|
582 |
+
#### Sentence Transformers
|
583 |
+
```bibtex
|
584 |
+
@inproceedings{reimers-2019-sentence-bert,
|
585 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
586 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
587 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
588 |
+
month = "11",
|
589 |
+
year = "2019",
|
590 |
+
publisher = "Association for Computational Linguistics",
|
591 |
+
url = "https://arxiv.org/abs/1908.10084",
|
592 |
+
}
|
593 |
+
```
|
594 |
+
|
595 |
+
#### MatryoshkaLoss
|
596 |
+
```bibtex
|
597 |
+
@misc{kusupati2024matryoshka,
|
598 |
+
title={Matryoshka Representation Learning},
|
599 |
+
author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi},
|
600 |
+
year={2024},
|
601 |
+
eprint={2205.13147},
|
602 |
+
archivePrefix={arXiv},
|
603 |
+
primaryClass={cs.LG}
|
604 |
+
}
|
605 |
+
```
|
606 |
+
|
607 |
+
#### MultipleNegativesRankingLoss
|
608 |
+
```bibtex
|
609 |
+
@misc{henderson2017efficient,
|
610 |
+
title={Efficient Natural Language Response Suggestion for Smart Reply},
|
611 |
+
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
|
612 |
+
year={2017},
|
613 |
+
eprint={1705.00652},
|
614 |
+
archivePrefix={arXiv},
|
615 |
+
primaryClass={cs.CL}
|
616 |
+
}
|
617 |
+
```
|
618 |
+
|
619 |
+
<!--
|
620 |
+
## Glossary
|
621 |
+
|
622 |
+
*Clearly define terms in order to be accessible across audiences.*
|
623 |
+
-->
|
624 |
+
|
625 |
+
<!--
|
626 |
+
## Model Card Authors
|
627 |
+
|
628 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
629 |
+
-->
|
630 |
+
|
631 |
+
<!--
|
632 |
+
## Model Card Contact
|
633 |
+
|
634 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
635 |
+
-->
|
config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Snowflake/snowflake-arctic-embed-l",
|
3 |
+
"architectures": [
|
4 |
+
"BertModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 1024,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 4096,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 16,
|
17 |
+
"num_hidden_layers": 24,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"torch_dtype": "float32",
|
21 |
+
"transformers_version": "4.49.0",
|
22 |
+
"type_vocab_size": 2,
|
23 |
+
"use_cache": true,
|
24 |
+
"vocab_size": 30522
|
25 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "3.4.1",
|
4 |
+
"transformers": "4.49.0",
|
5 |
+
"pytorch": "2.6.0"
|
6 |
+
},
|
7 |
+
"prompts": {
|
8 |
+
"query": "Represent this sentence for searching relevant passages: "
|
9 |
+
},
|
10 |
+
"default_prompt_name": null,
|
11 |
+
"similarity_fn_name": "cosine"
|
12 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4941569208f7c8800d39c3274f39f4d2c4710f5d224ded0d44da0a2977d42d3
|
3 |
+
size 1336413848
|
modules.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
},
|
14 |
+
{
|
15 |
+
"idx": 2,
|
16 |
+
"name": "2",
|
17 |
+
"path": "2_Normalize",
|
18 |
+
"type": "sentence_transformers.models.Normalize"
|
19 |
+
}
|
20 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 512,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"unk_token": {
|
31 |
+
"content": "[UNK]",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
}
|
37 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": true,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"max_length": 512,
|
50 |
+
"model_max_length": 512,
|
51 |
+
"pad_to_multiple_of": null,
|
52 |
+
"pad_token": "[PAD]",
|
53 |
+
"pad_token_type_id": 0,
|
54 |
+
"padding_side": "right",
|
55 |
+
"sep_token": "[SEP]",
|
56 |
+
"stride": 0,
|
57 |
+
"strip_accents": null,
|
58 |
+
"tokenize_chinese_chars": true,
|
59 |
+
"tokenizer_class": "BertTokenizer",
|
60 |
+
"truncation_side": "right",
|
61 |
+
"truncation_strategy": "longest_first",
|
62 |
+
"unk_token": "[UNK]"
|
63 |
+
}
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|