zihoo commited on
Commit
28b3247
·
verified ·
1 Parent(s): 4cee7df

Add new SentenceTransformer model.

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,540 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:160000
8
+ - loss:MarginDistillationLoss
9
+ base_model: sentence-transformers/all-MiniLM-L6-v2
10
+ widget:
11
+ - source_sentence: what is mindfulness at work
12
+ sentences:
13
+ - First, we define awareness as employees’ skills of being aware of their internal
14
+ experiences (e.g., mood and body) and the external environment (e.g., colleagues,
15
+ working relationships, and office environment) in work situations. Thus, the awareness
16
+ dimension incorporates both internal and external factors (Brown & Ryan, 2003).
17
+ Individuals have awareness when they are able to observe their own thoughts and
18
+ feelings as well as what is occurring in the environment. Thus, mindfulness entails
19
+ awareness of “phenomena occurring both externally and internally” (Dane, 2011,
20
+ p. 1000). This understanding is embedded in metacognitive skills, in the sense
21
+ that employees’ conscious return to mindfulness is an actionable practice that
22
+ involves recognizing that multiple response options exist for every stimulus encountered
23
+ at work (Dane & Brummel, 2014; Zivnuska et al., 2016). Prior research has suggested
24
+ that the basic feature of mindfulness is a metacognitive skill related to one’s
25
+ awareness—that is, objective awareness and following of internal experiences and
26
+ external events (Wells, 2002). As such, workplace mindfulness induces a metacognitive
27
+ mode of being fully aware of the current moment.
28
+ - 'A widely recognized and accepted view argues that mindfulness can be learned
29
+ and practiced as a set of skills through specific training (e.g., Baer et al.,
30
+ 2004, 2006; Fresco et al., 2007). For example, Baer et al. (2004) emphasized that
31
+ mindfulness skills are characterized by observing, describing, acting with awareness,
32
+ and accepting without judgment. Baer et al. (2006) indicated that mindfulness
33
+ encompasses five clear and interpretable facets of skills: nonreactivity, observing,
34
+ acting with awareness, describing, and nonjudging. Considering the malleability
35
+ of individuals at work, we adopt the skill view when developing our workplace
36
+ mindfulness scale to highlight that individual mindfulness at work can be trained
37
+ and improved. Indeed, individuals differ from one another in terms of their ability
38
+ to be mindful (Brown & Ryan, 2003; Kabat-Zinn, 1994). The changeability of mindfulness
39
+ underlines that this skill can be learned, trained, and improved through mindfulness
40
+ practices (Bishop et al., 2004; Walach et al., 2006). Workplace mindfulness differs
41
+ from both personality, which is difficult to change in the short term, and an
42
+ individual state that experiences large fluctuations on a daily basis (Hülsheger
43
+ et al., 2013; Olafsen, 2017).'
44
+ - Scholars have developed several measures of mindfulness (Table 1). These measures
45
+ help us understand the construct of mindfulness, but they are very different in
46
+ terms of conceptualization, factor structure, scoring, reliability, and validity.
47
+ For example, the Freiburg Mindfulness Inventory (FMI; Buchheld et al., 2001) and
48
+ Toronto Mindfulness Scale (TMS; Lau et al., 2006) were developed to measure states
49
+ of mindfulness. The Mindfulness Attention and Awareness Scale (MAAS; Brown & Ryan,
50
+ 2003), Cognitive and Affective Mindfulness Scale—Revised (CAMS-R; Feldman et al.,
51
+ 2007), and Philadelphia Mindfulness Questionnaire (PMQ; Cardaciotto et al., 2008)
52
+ have been employed to measure mindfulness as a trait. The Five Facet Mindfulness
53
+ Questionnaire (FFMQ; Baer et al., 2006), Experiences Questionnaire (EQ; Fresco
54
+ et al., 2007), and Kentucky Inventory of Mindfulness Skills (KIMS; Baer et al.,
55
+ 2004) seek to measure mindfulness skills. The Southampton Mindfulness Questionnaire
56
+ (SMQ; Chadwick et al., 2008) and Mindfulness/Mindlessness Scale (MMS; Haigh et
57
+ al., 2011) are intended to measure mindfulness as a cognitive process.
58
+ - source_sentence: define mindful attention
59
+ sentences:
60
+ - Second, we define attention as the skills of employees to focus their attention
61
+ on the present moment in work situations, and effectively turn their attention
62
+ back to the present in the face of interference. Sustained attention to the present
63
+ experience has been widely acknowledged as a key factor in mindfulness research
64
+ (Bishop et al., 2004). As a heightened sensitivity to a restricted range of experience,
65
+ attention implies that individuals pay full attention to the present and are not
66
+ susceptible to any distractions; even when they encounter interference, they do
67
+ not take too long to address it or have difficulty refocusing. Indeed, self-regulation
68
+ of attention allows for increased recognition of mental events in the present
69
+ moment (Bishop et al., 2004), which is also in line with the notion of metacognitive
70
+ skills—that is, mindful employees have the ability to intentionally regulate and
71
+ control their attention (Baer, 2003; Jankowski & Holas, 2014). This skill helps
72
+ employees not only maintain sustained attention, but also return from distractions
73
+ caused by irrelevant stimuli (Bishop et al., 2004; Teasdale, 1999). In essence,
74
+ it prevents their minds from wandering away from the present situation to potentially
75
+ negative conceptualizing (Kudesia, 2019; Schooler et al., 2011). Supporting this
76
+ notion, Fernandez-Duque et al. (2000) have stated that conscious regulation of
77
+ and control for attention directly refers to metacognitive skills.
78
+ - Supporting this rationale, Kudesia (2019) and Kudesia and Nyima (2015) highlighted
79
+ the context-specific nature of mindfulness, recognizing that mindfulness reflects
80
+ the individual’s metacognitive skills, which are engaged in a particular situation.
81
+ Bishop (2002) suggested that mindfulness practice actually encompasses several
82
+ metacognitive processes, proposing that mindfulness can be described as a type
83
+ of metacognitive skill. In essence, the metacognitive skills embedded in workplace
84
+ mindfulness are a crucial part of the self-regulatory loop, and their core purpose
85
+ is to reduce the discrepancy between the actual state and an undesired state,
86
+ or between an actual state and a desired state (Carver & Scheier, 2002). As such,
87
+ workplace mindfulness comprises employees’ metacognitive skills within the work
88
+ setting through which they demonstrate their regulation ability when dealing with
89
+ work-related activities. We argue that the individual’s specific skills become
90
+ especially pertinent when we consider particular aspects of the metacognitive
91
+ practice pathway of mindfulness, such as self-regulating attention and noticing
92
+ subtle affective sensations. For example, mindfulness is associated with the ability
93
+ to distance oneself from stimuli (Chambers et al., 2009). This distancing allows
94
+ employees to notice, prioritize, and respond to distractions in a conscious way,
95
+ without impulsivity or defensiveness (Bishop et al., 2004; Teasdale, 1999; Teasdale
96
+ et al., 1995).
97
+ - 'A widely recognized and accepted view argues that mindfulness can be learned
98
+ and practiced as a set of skills through specific training (e.g., Baer et al.,
99
+ 2004, 2006; Fresco et al., 2007). For example, Baer et al. (2004) emphasized that
100
+ mindfulness skills are characterized by observing, describing, acting with awareness,
101
+ and accepting without judgment. Baer et al. (2006) indicated that mindfulness
102
+ encompasses five clear and interpretable facets of skills: nonreactivity, observing,
103
+ acting with awareness, describing, and nonjudging. Considering the malleability
104
+ of individuals at work, we adopt the skill view when developing our workplace
105
+ mindfulness scale to highlight that individual mindfulness at work can be trained
106
+ and improved. Indeed, individuals differ from one another in terms of their ability
107
+ to be mindful (Brown & Ryan, 2003; Kabat-Zinn, 1994). The changeability of mindfulness
108
+ underlines that this skill can be learned, trained, and improved through mindfulness
109
+ practices (Bishop et al., 2004; Walach et al., 2006). Workplace mindfulness differs
110
+ from both personality, which is difficult to change in the short term, and an
111
+ individual state that experiences large fluctuations on a daily basis (Hülsheger
112
+ et al., 2013; Olafsen, 2017).'
113
+ - source_sentence: which definition of mindfulness is based on the cognitive process?
114
+ sentences:
115
+ - In the information processing framework, mindfulness is regarded as a cognitive
116
+ process (Chadwick et al., 2008; Haigh et al., 2011; Langer, 1989). According to
117
+ Langer (1989, p. 4), mindfulness is “a general style or mode of functioning through
118
+ which the individual actively engages in reconstructing the environment through
119
+ creating new categories or distinctions, thus directing attention to new contextual
120
+ cues that may be consciously controlled or manipulated as appropriate.” She proposed
121
+ that mindfulness includes novelty seeking, engagement, novelty producing, and
122
+ flexibility. The first two aspects refer to one’s orientation to the environment;
123
+ the latter two refer to how an individual operates within the environment (Bodner
124
+ & Langer, 2001).
125
+ - We adopt the view of mindfulness as a set of skills (Baer et al., 2004, 2006;
126
+ Fresco et al., 2007). The practice of mindfulness meditation highlights the necessity
127
+ of conceptualizing mindfulness as a skill that can be cultivated (Baer et al.,
128
+ 2004). To be specific, mindfulness can be a metacognitive practice that entails
129
+ monitoring and adjusting one’s information processing (Fernandez-Duque et al.,
130
+ 2000; Kudesia, 2019; Nelson, 1996); it can be improved by the fundamental meditation
131
+ technique (Kabat-Zinn, 1994). Mindfulness training allows individuals to become
132
+ aware of their internal feelings and external stimuli without reacting to them
133
+ (Hofmann et al., 2010). In addition to allowing for mindfulness interventions,
134
+ a skill-based conceptualization of workplace mindfulness is sensitive to other
135
+ external factors, such as the working environment (Kudesia, 2019; Kudesia & Nyima,
136
+ 2015). Indeed, skills should be considered context-specific because they have
137
+ varied expressions in different situations (Attewell, 1990), reflecting the focal
138
+ individual’s behavioral reactions at work toward unique experiences, events, and
139
+ environments. This view aligns with our definition and focus—that is, the workplace
140
+ situation as a specific context. We also emphasize that in a working system, we
141
+ cannot assess a person’s mindfulness without considering the people and the external
142
+ environment surrounding that individual.
143
+ - First, we define awareness as employees’ skills of being aware of their internal
144
+ experiences (e.g., mood and body) and the external environment (e.g., colleagues,
145
+ working relationships, and office environment) in work situations. Thus, the awareness
146
+ dimension incorporates both internal and external factors (Brown & Ryan, 2003).
147
+ Individuals have awareness when they are able to observe their own thoughts and
148
+ feelings as well as what is occurring in the environment. Thus, mindfulness entails
149
+ awareness of “phenomena occurring both externally and internally” (Dane, 2011,
150
+ p. 1000). This understanding is embedded in metacognitive skills, in the sense
151
+ that employees’ conscious return to mindfulness is an actionable practice that
152
+ involves recognizing that multiple response options exist for every stimulus encountered
153
+ at work (Dane & Brummel, 2014; Zivnuska et al., 2016). Prior research has suggested
154
+ that the basic feature of mindfulness is a metacognitive skill related to one’s
155
+ awareness—that is, objective awareness and following of internal experiences and
156
+ external events (Wells, 2002). As such, workplace mindfulness induces a metacognitive
157
+ mode of being fully aware of the current moment.
158
+ - source_sentence: what is acceptance in mindfulness
159
+ sentences:
160
+ - 'Mindfulness is widely considered as “paying attention in a particular way: on
161
+ purpose, in the present moment, and nonjudgmentally” (Kabat-Zinn, 1994, p. 4).
162
+ However, scholars have not reached a consensus on the essential features of mindfulness,
163
+ with various conceptualizations such as a set of skills, a state, a trait, and
164
+ a cognitive process. In what follows, we summarize the prevailing views of mindfulness
165
+ in the literature.'
166
+ - Second, we define attention as the skills of employees to focus their attention
167
+ on the present moment in work situations, and effectively turn their attention
168
+ back to the present in the face of interference. Sustained attention to the present
169
+ experience has been widely acknowledged as a key factor in mindfulness research
170
+ (Bishop et al., 2004). As a heightened sensitivity to a restricted range of experience,
171
+ attention implies that individuals pay full attention to the present and are not
172
+ susceptible to any distractions; even when they encounter interference, they do
173
+ not take too long to address it or have difficulty refocusing. Indeed, self-regulation
174
+ of attention allows for increased recognition of mental events in the present
175
+ moment (Bishop et al., 2004), which is also in line with the notion of metacognitive
176
+ skills—that is, mindful employees have the ability to intentionally regulate and
177
+ control their attention (Baer, 2003; Jankowski & Holas, 2014). This skill helps
178
+ employees not only maintain sustained attention, but also return from distractions
179
+ caused by irrelevant stimuli (Bishop et al., 2004; Teasdale, 1999). In essence,
180
+ it prevents their minds from wandering away from the present situation to potentially
181
+ negative conceptualizing (Kudesia, 2019; Schooler et al., 2011). Supporting this
182
+ notion, Fernandez-Duque et al. (2000) have stated that conscious regulation of
183
+ and control for attention directly refers to metacognitive skills.
184
+ - Brown and Ryan (2003, 2004) also suggest that acceptance, which is characterized
185
+ by openness or receptivity to experiences and events, is a key component of mindfulness
186
+ and is not redundant with awareness. Although they consider this dimension to
187
+ be subsumed within the individual’s capacity to sustain attention to and remain
188
+ aware of the present moment, acceptance remains a useful facet of mindfulness
189
+ to address. It reflects the individual’s acceptance of internal and external phenomena
190
+ without judging them (Baer, 2003). Research has also supported the recognition
191
+ of acceptance as a dimension of mindfulness (Baer et al., 2004, 2006; Buchheld
192
+ et al., 2001).
193
+ - source_sentence: what is acceptance in mindfulness
194
+ sentences:
195
+ - Second, we define attention as the skills of employees to focus their attention
196
+ on the present moment in work situations, and effectively turn their attention
197
+ back to the present in the face of interference. Sustained attention to the present
198
+ experience has been widely acknowledged as a key factor in mindfulness research
199
+ (Bishop et al., 2004). As a heightened sensitivity to a restricted range of experience,
200
+ attention implies that individuals pay full attention to the present and are not
201
+ susceptible to any distractions; even when they encounter interference, they do
202
+ not take too long to address it or have difficulty refocusing. Indeed, self-regulation
203
+ of attention allows for increased recognition of mental events in the present
204
+ moment (Bishop et al., 2004), which is also in line with the notion of metacognitive
205
+ skills—that is, mindful employees have the ability to intentionally regulate and
206
+ control their attention (Baer, 2003; Jankowski & Holas, 2014). This skill helps
207
+ employees not only maintain sustained attention, but also return from distractions
208
+ caused by irrelevant stimuli (Bishop et al., 2004; Teasdale, 1999). In essence,
209
+ it prevents their minds from wandering away from the present situation to potentially
210
+ negative conceptualizing (Kudesia, 2019; Schooler et al., 2011). Supporting this
211
+ notion, Fernandez-Duque et al. (2000) have stated that conscious regulation of
212
+ and control for attention directly refers to metacognitive skills.
213
+ - Brown and Ryan (2003, 2004) also suggest that acceptance, which is characterized
214
+ by openness or receptivity to experiences and events, is a key component of mindfulness
215
+ and is not redundant with awareness. Although they consider this dimension to
216
+ be subsumed within the individual’s capacity to sustain attention to and remain
217
+ aware of the present moment, acceptance remains a useful facet of mindfulness
218
+ to address. It reflects the individual’s acceptance of internal and external phenomena
219
+ without judging them (Baer, 2003). Research has also supported the recognition
220
+ of acceptance as a dimension of mindfulness (Baer et al., 2004, 2006; Buchheld
221
+ et al., 2001).
222
+ - Brown and Ryan (2003) further propose that, despite their intertwined nature,
223
+ distinctions exist between attention and awareness—the insights gained by sustained
224
+ awareness can only be translated into specific actions by paying focused attention
225
+ to our behaviors or the tasks at hand (Martin, 1997). Hence, heightened attention
226
+ to and awareness of experiences and events should capture two different aspects
227
+ of mindfulness. Recent research has also emphasized that attention and awareness
228
+ should be distinguished from each other because attention reflects an ever-changing
229
+ factor of consciousness, whereas awareness refers to a specific and stable state
230
+ of consciousness (Selart et al., in press). In the past, attention and awareness
231
+ have proved important to the study of mindfulness-promoting practices (Brown &
232
+ Ryan, 2004), as some of these practices highlight focused attention whereas others
233
+ emphasize awareness (Bishop et al., 2004). Notably, research has yielded empirical
234
+ support confirming these distinctions (Feldman et al., 2007).
235
+ pipeline_tag: sentence-similarity
236
+ library_name: sentence-transformers
237
+ ---
238
+
239
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
240
+
241
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
242
+
243
+ ## Model Details
244
+
245
+ ### Model Description
246
+ - **Model Type:** Sentence Transformer
247
+ - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision fa97f6e7cb1a59073dff9e6b13e2715cf7475ac9 -->
248
+ - **Maximum Sequence Length:** 350 tokens
249
+ - **Output Dimensionality:** 384 dimensions
250
+ - **Similarity Function:** Cosine Similarity
251
+ <!-- - **Training Dataset:** Unknown -->
252
+ <!-- - **Language:** Unknown -->
253
+ <!-- - **License:** Unknown -->
254
+
255
+ ### Model Sources
256
+
257
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
258
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
259
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
260
+
261
+ ### Full Model Architecture
262
+
263
+ ```
264
+ SentenceTransformer(
265
+ (0): Transformer({'max_seq_length': 350, 'do_lower_case': False}) with Transformer model: BertModel
266
+ (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
267
+ (2): Normalize()
268
+ )
269
+ ```
270
+
271
+ ## Usage
272
+
273
+ ### Direct Usage (Sentence Transformers)
274
+
275
+ First install the Sentence Transformers library:
276
+
277
+ ```bash
278
+ pip install -U sentence-transformers
279
+ ```
280
+
281
+ Then you can load this model and run inference.
282
+ ```python
283
+ from sentence_transformers import SentenceTransformer
284
+
285
+ # Download from the 🤗 Hub
286
+ model = SentenceTransformer("zihoo/all-MiniLM-L6-v2-WMGPL")
287
+ # Run inference
288
+ sentences = [
289
+ 'what is acceptance in mindfulness',
290
+ 'Brown and Ryan (2003, 2004) also suggest that acceptance, which is characterized by openness or receptivity to experiences and events, is a key component of mindfulness and is not redundant with awareness. Although they consider this dimension to be subsumed within the individual’s capacity to sustain attention to and remain aware of the present moment, acceptance remains a useful facet of mindfulness to address. It reflects the individual’s acceptance of internal and external phenomena without judging them (Baer, 2003). Research has also supported the recognition of acceptance as a dimension of mindfulness (Baer et al., 2004, 2006; Buchheld et al., 2001).',
291
+ 'Brown and Ryan (2003) further propose that, despite their intertwined nature, distinctions exist between attention and awareness—the insights gained by sustained awareness can only be translated into specific actions by paying focused attention to our behaviors or the tasks at hand (Martin, 1997). Hence, heightened attention to and awareness of experiences and events should capture two different aspects of mindfulness. Recent research has also emphasized that attention and awareness should be distinguished from each other because attention reflects an ever-changing factor of consciousness, whereas awareness refers to a specific and stable state of consciousness (Selart et al., in press). In the past, attention and awareness have proved important to the study of mindfulness-promoting practices (Brown & Ryan, 2004), as some of these practices highlight focused attention whereas others emphasize awareness (Bishop et al., 2004). Notably, research has yielded empirical support confirming these distinctions (Feldman et al., 2007).',
292
+ ]
293
+ embeddings = model.encode(sentences)
294
+ print(embeddings.shape)
295
+ # [3, 384]
296
+
297
+ # Get the similarity scores for the embeddings
298
+ similarities = model.similarity(embeddings, embeddings)
299
+ print(similarities.shape)
300
+ # [3, 3]
301
+ ```
302
+
303
+ <!--
304
+ ### Direct Usage (Transformers)
305
+
306
+ <details><summary>Click to see the direct usage in Transformers</summary>
307
+
308
+ </details>
309
+ -->
310
+
311
+ <!--
312
+ ### Downstream Usage (Sentence Transformers)
313
+
314
+ You can finetune this model on your own dataset.
315
+
316
+ <details><summary>Click to expand</summary>
317
+
318
+ </details>
319
+ -->
320
+
321
+ <!--
322
+ ### Out-of-Scope Use
323
+
324
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
325
+ -->
326
+
327
+ <!--
328
+ ## Bias, Risks and Limitations
329
+
330
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
331
+ -->
332
+
333
+ <!--
334
+ ### Recommendations
335
+
336
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
337
+ -->
338
+
339
+ ## Training Details
340
+
341
+ ### Training Dataset
342
+
343
+ #### Unnamed Dataset
344
+
345
+
346
+ * Size: 160,000 training samples
347
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, <code>sentence_2</code>, and <code>label</code>
348
+ * Approximate statistics based on the first 1000 samples:
349
+ | | sentence_0 | sentence_1 | sentence_2 | label |
350
+ |:--------|:---------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------|
351
+ | type | string | string | string | float |
352
+ | details | <ul><li>min: 5 tokens</li><li>mean: 8.74 tokens</li><li>max: 14 tokens</li></ul> | <ul><li>min: 102 tokens</li><li>mean: 234.76 tokens</li><li>max: 350 tokens</li></ul> | <ul><li>min: 102 tokens</li><li>mean: 232.57 tokens</li><li>max: 350 tokens</li></ul> | <ul><li>min: -6.82</li><li>mean: 5.05</li><li>max: 19.99</li></ul> |
353
+ * Samples:
354
+ | sentence_0 | sentence_1 | sentence_2 | label |
355
+ |:-------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------|
356
+ | <code>which dimensions of mindfulness are useful in the workplace</code> | <code>While Brown and Ryan (2003, 2004) highlight the three key components of mindfulness and subsequent research has largely agreed upon this framework (Bishop et al., 2004), existing scales—including the one developed by Brown and Ryan (2003)—do not actually include all three components. Only by capturing the entire set of dimensions of workplace mindfulness can we fully understand this construct and examine its applicability in the workplace. Notably, although other dimensions have been examined in existing scales, some of them are not primarily relevant to the work context. More importantly, this research does not adopt an all-encompassing approach, but rather draws on just Brown and Ryan’s theory (2003, 2004) to develop the dimensions.</code> | <code>Mindfulness is also defined as a state (e.g., Bishop et al., 2004; Good et al., 2016; Lau et al., 2006; Tanay & Bernstein, 2013) of being aware of and attentive to what is taking place internally and externally at that moment (Good et al., 2016; Lau et al., 2006; Tanay & Bernstein, 2013). For example, Lau et al., (2006, p. 1447) described mindfulness as “a mode, or state-like quality that is maintained only when attention to experience is intentionally cultivated with an open, nonjudgmental orientation to experience.” More recently, Good et al., (2016, p. 117) defined mindfulness as “receptive attention to and awareness of present events and experience.</code> | <code>7.988612174987793</code> |
357
+ | <code>how do we understand mindfulness in the workplace?</code> | <code>While Brown and Ryan (2003, 2004) highlight the three key components of mindfulness and subsequent research has largely agreed upon this framework (Bishop et al., 2004), existing scales—including the one developed by Brown and Ryan (2003)—do not actually include all three components. Only by capturing the entire set of dimensions of workplace mindfulness can we fully understand this construct and examine its applicability in the workplace. Notably, although other dimensions have been examined in existing scales, some of them are not primarily relevant to the work context. More importantly, this research does not adopt an all-encompassing approach, but rather draws on just Brown and Ryan’s theory (2003, 2004) to develop the dimensions.</code> | <code>In the information processing framework, mindfulness is regarded as a cognitive process (Chadwick et al., 2008; Haigh et al., 2011; Langer, 1989). According to Langer (1989, p. 4), mindfulness is “a general style or mode of functioning through which the individual actively engages in reconstructing the environment through creating new categories or distinctions, thus directing attention to new contextual cues that may be consciously controlled or manipulated as appropriate.” She proposed that mindfulness includes novelty seeking, engagement, novelty producing, and flexibility. The first two aspects refer to one’s orientation to the environment; the latter two refer to how an individual operates within the environment (Bodner & Langer, 2001).</code> | <code>3.691136598587036</code> |
358
+ | <code>what does mindfulness mean</code> | <code>Scholars have developed several measures of mindfulness (Table 1). These measures help us understand the construct of mindfulness, but they are very different in terms of conceptualization, factor structure, scoring, reliability, and validity. For example, the Freiburg Mindfulness Inventory (FMI; Buchheld et al., 2001) and Toronto Mindfulness Scale (TMS; Lau et al., 2006) were developed to measure states of mindfulness. The Mindfulness Attention and Awareness Scale (MAAS; Brown & Ryan, 2003), Cognitive and Affective Mindfulness Scale—Revised (CAMS-R; Feldman et al., 2007), and Philadelphia Mindfulness Questionnaire (PMQ; Cardaciotto et al., 2008) have been employed to measure mindfulness as a trait. The Five Facet Mindfulness Questionnaire (FFMQ; Baer et al., 2006), Experiences Questionnaire (EQ; Fresco et al., 2007), and Kentucky Inventory of Mindfulness Skills (KIMS; Baer et al., 2004) seek to measure mindfulness skills. The Southampton Mindfulness Questionnaire (SMQ; Chadwick et al....</code> | <code>Brown and Ryan (2003) further propose that, despite their intertwined nature, distinctions exist between attention and awareness—the insights gained by sustained awareness can only be translated into specific actions by paying focused attention to our behaviors or the tasks at hand (Martin, 1997). Hence, heightened attention to and awareness of experiences and events should capture two different aspects of mindfulness. Recent research has also emphasized that attention and awareness should be distinguished from each other because attention reflects an ever-changing factor of consciousness, whereas awareness refers to a specific and stable state of consciousness (Selart et al., in press). In the past, attention and awareness have proved important to the study of mindfulness-promoting practices (Brown & Ryan, 2004), as some of these practices highlight focused attention whereas others emphasize awareness (Bishop et al., 2004). Notably, research has yielded empirical support confirming th...</code> | <code>0.38913965225219727</code> |
359
+ * Loss: <code>gpl.toolkit.loss.MarginDistillationLoss</code>
360
+
361
+ ### Training Hyperparameters
362
+ #### Non-Default Hyperparameters
363
+
364
+ - `per_device_train_batch_size`: 16
365
+ - `per_device_eval_batch_size`: 16
366
+ - `num_train_epochs`: 1
367
+ - `max_steps`: 1000
368
+ - `multi_dataset_batch_sampler`: round_robin
369
+
370
+ #### All Hyperparameters
371
+ <details><summary>Click to expand</summary>
372
+
373
+ - `overwrite_output_dir`: False
374
+ - `do_predict`: False
375
+ - `eval_strategy`: no
376
+ - `prediction_loss_only`: True
377
+ - `per_device_train_batch_size`: 16
378
+ - `per_device_eval_batch_size`: 16
379
+ - `per_gpu_train_batch_size`: None
380
+ - `per_gpu_eval_batch_size`: None
381
+ - `gradient_accumulation_steps`: 1
382
+ - `eval_accumulation_steps`: None
383
+ - `torch_empty_cache_steps`: None
384
+ - `learning_rate`: 5e-05
385
+ - `weight_decay`: 0.0
386
+ - `adam_beta1`: 0.9
387
+ - `adam_beta2`: 0.999
388
+ - `adam_epsilon`: 1e-08
389
+ - `max_grad_norm`: 1
390
+ - `num_train_epochs`: 1
391
+ - `max_steps`: 1000
392
+ - `lr_scheduler_type`: linear
393
+ - `lr_scheduler_kwargs`: {}
394
+ - `warmup_ratio`: 0.0
395
+ - `warmup_steps`: 0
396
+ - `log_level`: passive
397
+ - `log_level_replica`: warning
398
+ - `log_on_each_node`: True
399
+ - `logging_nan_inf_filter`: True
400
+ - `save_safetensors`: True
401
+ - `save_on_each_node`: False
402
+ - `save_only_model`: False
403
+ - `restore_callback_states_from_checkpoint`: False
404
+ - `no_cuda`: False
405
+ - `use_cpu`: False
406
+ - `use_mps_device`: False
407
+ - `seed`: 42
408
+ - `data_seed`: None
409
+ - `jit_mode_eval`: False
410
+ - `use_ipex`: False
411
+ - `bf16`: False
412
+ - `fp16`: False
413
+ - `fp16_opt_level`: O1
414
+ - `half_precision_backend`: auto
415
+ - `bf16_full_eval`: False
416
+ - `fp16_full_eval`: False
417
+ - `tf32`: None
418
+ - `local_rank`: 0
419
+ - `ddp_backend`: None
420
+ - `tpu_num_cores`: None
421
+ - `tpu_metrics_debug`: False
422
+ - `debug`: []
423
+ - `dataloader_drop_last`: False
424
+ - `dataloader_num_workers`: 0
425
+ - `dataloader_prefetch_factor`: None
426
+ - `past_index`: -1
427
+ - `disable_tqdm`: False
428
+ - `remove_unused_columns`: True
429
+ - `label_names`: None
430
+ - `load_best_model_at_end`: False
431
+ - `ignore_data_skip`: False
432
+ - `fsdp`: []
433
+ - `fsdp_min_num_params`: 0
434
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
435
+ - `fsdp_transformer_layer_cls_to_wrap`: None
436
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
437
+ - `deepspeed`: None
438
+ - `label_smoothing_factor`: 0.0
439
+ - `optim`: adamw_torch
440
+ - `optim_args`: None
441
+ - `adafactor`: False
442
+ - `group_by_length`: False
443
+ - `length_column_name`: length
444
+ - `ddp_find_unused_parameters`: None
445
+ - `ddp_bucket_cap_mb`: None
446
+ - `ddp_broadcast_buffers`: False
447
+ - `dataloader_pin_memory`: True
448
+ - `dataloader_persistent_workers`: False
449
+ - `skip_memory_metrics`: True
450
+ - `use_legacy_prediction_loop`: False
451
+ - `push_to_hub`: False
452
+ - `resume_from_checkpoint`: None
453
+ - `hub_model_id`: None
454
+ - `hub_strategy`: every_save
455
+ - `hub_private_repo`: None
456
+ - `hub_always_push`: False
457
+ - `gradient_checkpointing`: False
458
+ - `gradient_checkpointing_kwargs`: None
459
+ - `include_inputs_for_metrics`: False
460
+ - `include_for_metrics`: []
461
+ - `eval_do_concat_batches`: True
462
+ - `fp16_backend`: auto
463
+ - `push_to_hub_model_id`: None
464
+ - `push_to_hub_organization`: None
465
+ - `mp_parameters`:
466
+ - `auto_find_batch_size`: False
467
+ - `full_determinism`: False
468
+ - `torchdynamo`: None
469
+ - `ray_scope`: last
470
+ - `ddp_timeout`: 1800
471
+ - `torch_compile`: False
472
+ - `torch_compile_backend`: None
473
+ - `torch_compile_mode`: None
474
+ - `dispatch_batches`: None
475
+ - `split_batches`: None
476
+ - `include_tokens_per_second`: False
477
+ - `include_num_input_tokens_seen`: False
478
+ - `neftune_noise_alpha`: None
479
+ - `optim_target_modules`: None
480
+ - `batch_eval_metrics`: False
481
+ - `eval_on_start`: False
482
+ - `use_liger_kernel`: False
483
+ - `eval_use_gather_object`: False
484
+ - `average_tokens_across_devices`: False
485
+ - `prompts`: None
486
+ - `batch_sampler`: batch_sampler
487
+ - `multi_dataset_batch_sampler`: round_robin
488
+
489
+ </details>
490
+
491
+ ### Training Logs
492
+ | Epoch | Step | Training Loss |
493
+ |:-----:|:----:|:-------------:|
494
+ | 0.05 | 500 | 44.9354 |
495
+ | 0.1 | 1000 | 41.9204 |
496
+
497
+
498
+ ### Framework Versions
499
+ - Python: 3.11.11
500
+ - Sentence Transformers: 3.3.1
501
+ - Transformers: 4.47.1
502
+ - PyTorch: 2.5.1+cu121
503
+ - Accelerate: 1.2.1
504
+ - Datasets: 3.2.0
505
+ - Tokenizers: 0.21.0
506
+
507
+ ## Citation
508
+
509
+ ### BibTeX
510
+
511
+ #### Sentence Transformers
512
+ ```bibtex
513
+ @inproceedings{reimers-2019-sentence-bert,
514
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
515
+ author = "Reimers, Nils and Gurevych, Iryna",
516
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
517
+ month = "11",
518
+ year = "2019",
519
+ publisher = "Association for Computational Linguistics",
520
+ url = "https://arxiv.org/abs/1908.10084",
521
+ }
522
+ ```
523
+
524
+ <!--
525
+ ## Glossary
526
+
527
+ *Clearly define terms in order to be accessible across audiences.*
528
+ -->
529
+
530
+ <!--
531
+ ## Model Card Authors
532
+
533
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
534
+ -->
535
+
536
+ <!--
537
+ ## Model Card Contact
538
+
539
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
540
+ -->
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./output/WM_model",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.47.1",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.3.1",
4
+ "transformers": "4.47.1",
5
+ "pytorch": "2.5.1+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bb605f19cf448a4322540993221f3180386193ad66ecf3f36cf8df8ef93b44c
3
+ size 90864192
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 350,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 128,
51
+ "model_max_length": 350,
52
+ "never_split": null,
53
+ "pad_to_multiple_of": null,
54
+ "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
+ "sep_token": "[SEP]",
58
+ "stride": 0,
59
+ "strip_accents": null,
60
+ "tokenize_chinese_chars": true,
61
+ "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]"
65
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff