Add some metadata to the model card
#1
by
tomaarsen
HF Staff
- opened
README.md
CHANGED
@@ -8,6 +8,9 @@ language:
|
|
8 |
- es
|
9 |
base_model:
|
10 |
- PleIAs/Pleias-350m-Preview
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
|
@@ -91,4 +94,4 @@ With only 350 million parameters, Pleias-RAG-350M is classified among the *phone
|
|
91 |
|
92 |
We also release an unquantized [GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-350M-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
93 |
|
94 |
-
Once integrated into a RAG system, Pleias-RAG-350M can also be use in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make tiny model workable in production by relying systematically on an externalized memory.
|
|
|
8 |
- es
|
9 |
base_model:
|
10 |
- PleIAs/Pleias-350m-Preview
|
11 |
+
pipeline_tag: text-generation
|
12 |
+
tags:
|
13 |
+
- transformers
|
14 |
---
|
15 |
|
16 |
|
|
|
94 |
|
95 |
We also release an unquantized [GGUF version](https://huggingface.co/PleIAs/Pleias-RAG-350M-gguf) for deployment on CPU. Our internal performance benchmarks suggest that waiting times are currently acceptable for most either even under constrained RAM: about 20 seconds for a complex generation including reasoning traces on 8g RAM and below. Since the model is unquantized, quality of text generation should be identical to the original model.
|
96 |
|
97 |
+
Once integrated into a RAG system, Pleias-RAG-350M can also be use in a broader range of non-conversational use cases including user support or educational assistance. Through this release, we aims to make tiny model workable in production by relying systematically on an externalized memory.
|