Update README.md
Browse files
README.md
CHANGED
@@ -36,9 +36,8 @@ tags:
|
|
36 |
|
37 |
## Overview
|
38 |
|
39 |
-
EuroBERT is a family of
|
40 |
-
|
41 |
-
EuroBERT incorporates recent architectural advances from decoder models (such as RoPE and Flash Attention) and is trained on a 5T-token multilingual dataset covering 8 European languages and 7 widely spoken global languages, along with mathematics and code. EuroBERT models exhibit the strongest multilingual performance across domains and tasks compared to similarly sized systems.
|
42 |
|
43 |
It is available in the following sizes:
|
44 |
|
@@ -46,7 +45,7 @@ It is available in the following sizes:
|
|
46 |
- [EuroBERT-610m](https://huggingface.co/EuroBERT/EuroBERT-610m) - 26 layers, 610 million parameters
|
47 |
- [EuroBERT-2.1B](https://huggingface.co/EuroBERT/EuroBERT-2.1B) - 32 layers, 2.1 billion parameters
|
48 |
|
49 |
-
For more information about EuroBERT, please refer to the [release blog post](
|
50 |
|
51 |
## Usage
|
52 |
|
|
|
36 |
|
37 |
## Overview
|
38 |
|
39 |
+
EuroBERT is a family of multilingual encoder models designed for a variety of tasks—such as classification, retrieval, or evaluation metrics— supporting 15 languages, mathematics and code, and sequences of up to 8,192 tokens.
|
40 |
+
EuroBERT models exhibit the strongest multilingual performance across [domains and tasks](#Evaluation) compared to similarly sized systems.
|
|
|
41 |
|
42 |
It is available in the following sizes:
|
43 |
|
|
|
45 |
- [EuroBERT-610m](https://huggingface.co/EuroBERT/EuroBERT-610m) - 26 layers, 610 million parameters
|
46 |
- [EuroBERT-2.1B](https://huggingface.co/EuroBERT/EuroBERT-2.1B) - 32 layers, 2.1 billion parameters
|
47 |
|
48 |
+
For more information about EuroBERT, please refer to the [release blog post](***) for a high-level overview and our [arXiv pre-print](***) for in-depth information.
|
49 |
|
50 |
## Usage
|
51 |
|