Fill-Mask
Transformers
PyTorch
eurobert
code
custom_code
Nicolas-BZRD commited on
Commit
b023e4a
·
verified ·
1 Parent(s): 472905d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -4
README.md CHANGED
@@ -36,9 +36,8 @@ tags:
36
 
37
  ## Overview
38
 
39
- EuroBERT is a family of the most powerful multilingual encoder foundation models designed for a variety of tasks—such as classification, retrieval, or evaluation metrics—across 15 languages, including mathematics and code, and supports sequences of up to 8,192 tokens.
40
-
41
- EuroBERT incorporates recent architectural advances from decoder models (such as RoPE and Flash Attention) and is trained on a 5T-token multilingual dataset covering 8 European languages and 7 widely spoken global languages, along with mathematics and code. EuroBERT models exhibit the strongest multilingual performance across domains and tasks compared to similarly sized systems.
42
 
43
  It is available in the following sizes:
44
 
@@ -46,7 +45,7 @@ It is available in the following sizes:
46
  - [EuroBERT-610m](https://huggingface.co/EuroBERT/EuroBERT-610m) - 26 layers, 610 million parameters
47
  - [EuroBERT-2.1B](https://huggingface.co/EuroBERT/EuroBERT-2.1B) - 32 layers, 2.1 billion parameters
48
 
49
- For more information about EuroBERT, please refer to the [release blog post](https://huggingface.co/blog/modernbert) for a high-level overview and our [arXiv pre-print](https://arxiv.org/abs/2412.13663) for in-depth information.
50
 
51
  ## Usage
52
 
 
36
 
37
  ## Overview
38
 
39
+ EuroBERT is a family of multilingual encoder models designed for a variety of tasks—such as classification, retrieval, or evaluation metrics— supporting 15 languages, mathematics and code, and sequences of up to 8,192 tokens.
40
+ EuroBERT models exhibit the strongest multilingual performance across [domains and tasks](#Evaluation) compared to similarly sized systems.
 
41
 
42
  It is available in the following sizes:
43
 
 
45
  - [EuroBERT-610m](https://huggingface.co/EuroBERT/EuroBERT-610m) - 26 layers, 610 million parameters
46
  - [EuroBERT-2.1B](https://huggingface.co/EuroBERT/EuroBERT-2.1B) - 32 layers, 2.1 billion parameters
47
 
48
+ For more information about EuroBERT, please refer to the [release blog post](***) for a high-level overview and our [arXiv pre-print](***) for in-depth information.
49
 
50
  ## Usage
51