zpn commited on
Commit
3b160e9
·
verified ·
1 Parent(s): 24132aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -1
README.md CHANGED
@@ -119,7 +119,7 @@ nomic-embed-text-v2-moe is SoTA multilingual MoE text embedding model:
119
  - **High Performance**: SoTA Multilingual performance compared to ~300M parameter models, competitive with models 2x in size
120
  - **Multilinguality**: Supports ~100 languages and trained over 1.6B pairs
121
  - **Flexible Embedding Dimension**: Trained with [Matryoshka Embeddings](https://arxiv.org/abs/2205.13147) with 3x reductions in storage cost with minimal performance degredations
122
- - **Fully-Open Source**: Model weights, [code](https://github.com/nomic-ai/contrastors), and training data (see code repo) released
123
 
124
 
125
  | Model | Params (M) | Emb Dim | BEIR | MIRACL | Pretrain Data | Finetune Data | Code |
@@ -151,6 +151,12 @@ nomic-embed-text-v2-moe is SoTA multilingual MoE text embedding model:
151
 
152
  The model can be used through SentenceTransformers and Transformers.
153
 
 
 
 
 
 
 
154
  **Important**: the text prompt *must* include a *task instruction prefix*, instructing the model which task is being performed.
155
 
156
  For queries/questions, please use `search_query: ` and `search_document: ` for the corresponding document
 
119
  - **High Performance**: SoTA Multilingual performance compared to ~300M parameter models, competitive with models 2x in size
120
  - **Multilinguality**: Supports ~100 languages and trained over 1.6B pairs
121
  - **Flexible Embedding Dimension**: Trained with [Matryoshka Embeddings](https://arxiv.org/abs/2205.13147) with 3x reductions in storage cost with minimal performance degredations
122
+ - **Fully Open-Source**: Model weights, [code](https://github.com/nomic-ai/contrastors), and training data (see code repo) released
123
 
124
 
125
  | Model | Params (M) | Emb Dim | BEIR | MIRACL | Pretrain Data | Finetune Data | Code |
 
151
 
152
  The model can be used through SentenceTransformers and Transformers.
153
 
154
+ For best performance on GPU, please install
155
+
156
+ ```bash
157
+ pip install torch transformers einops git+https://github.com/nomic-ai/megablocks.git
158
+ ```
159
+
160
  **Important**: the text prompt *must* include a *task instruction prefix*, instructing the model which task is being performed.
161
 
162
  For queries/questions, please use `search_query: ` and `search_document: ` for the corresponding document