Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,7 @@ widget:
|
|
6 |
- text: " السلام عليكم ورحمة[MASK] وبركاتة"
|
7 |
- example_title: "Example 1"
|
8 |
---
|
|
|
9 |
|
10 |
**AraBERTMo** is an Arabic pre-trained language model based on [Google's BERT architechture](https://github.com/google-research/bert). AraBERTMo uses the same BERT-Base config.
|
11 |
|
@@ -13,6 +14,14 @@ AraBERTMo now comes in 10 new variants
|
|
13 |
All models are available on the `HuggingFace` model page under the [Ebtihal](https://huggingface.co/Ebtihal/) name. Checkpoints are available in PyTorch formats.
|
14 |
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
## Load Pretrained Model
|
17 |
|
18 |
You can use this model by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this:
|
|
|
6 |
- text: " السلام عليكم ورحمة[MASK] وبركاتة"
|
7 |
- example_title: "Example 1"
|
8 |
---
|
9 |
+
# Arabic BERT Mini Model
|
10 |
|
11 |
**AraBERTMo** is an Arabic pre-trained language model based on [Google's BERT architechture](https://github.com/google-research/bert). AraBERTMo uses the same BERT-Base config.
|
12 |
|
|
|
14 |
All models are available on the `HuggingFace` model page under the [Ebtihal](https://huggingface.co/Ebtihal/) name. Checkpoints are available in PyTorch formats.
|
15 |
|
16 |
|
17 |
+
## Pretraining Corpus
|
18 |
+
|
19 |
+
`AraBertMo_base_V1' model was pretrained on ~3 million words:
|
20 |
+
|
21 |
+
- Arabic version of [OSCAR](https://traces1.inria.fr/oscar/) -
|
22 |
+
|
23 |
+
|
24 |
+
|
25 |
## Load Pretrained Model
|
26 |
|
27 |
You can use this model by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this:
|